diff options
Diffstat (limited to 'python/openvino/demo/models')
| -rw-r--r-- | python/openvino/demo/models/architectures/A10_Small_NoSoftmax.arch | 118 | ||||
| -rw-r--r-- | python/openvino/demo/models/architectures/AGX7_Small_NoSoftmax.arch | 118 | ||||
| -rw-r--r-- | python/openvino/demo/models/architectures/agx7_area.arch | 111 | ||||
| -rw-r--r-- | python/openvino/demo/models/architectures/agx7_area.ptc | 248 | ||||
| -rw-r--r-- | python/openvino/demo/models/architectures/agx7_performance.arch | 111 | ||||
| -rw-r--r-- | python/openvino/demo/models/architectures/agx7_performance.ptc | 223 | ||||
| -rw-r--r-- | python/openvino/demo/models/architectures/de5a_area.arch | 111 | ||||
| -rw-r--r-- | python/openvino/demo/models/architectures/de5a_performance.arch | 111 | ||||
| -rw-r--r-- | python/openvino/demo/models/architectures/quartus_ptc.log | 26 | ||||
| -rw-r--r-- | python/openvino/demo/models/architectures/quartus_ptc.rec | 10 | ||||
| -rw-r--r-- | python/openvino/demo/models/model_bmp.bin | bin | 0 -> 1949252 bytes | |||
| -rw-r--r-- | python/openvino/demo/models/model_bmp.xml | 1782 | ||||
| -rw-r--r-- | python/openvino/demo/models/model_bmp_de5a_area.bin | bin | 0 -> 3374709 bytes |
13 files changed, 2969 insertions, 0 deletions
diff --git a/python/openvino/demo/models/architectures/A10_Small_NoSoftmax.arch b/python/openvino/demo/models/architectures/A10_Small_NoSoftmax.arch new file mode 100644 index 0000000..183d50d --- /dev/null +++ b/python/openvino/demo/models/architectures/A10_Small_NoSoftmax.arch @@ -0,0 +1,118 @@ +family : 'A10' +k_vector : 16 +c_vector : 16 +arch_precision : FP11 + +stream_buffer_depth : 17600 + +output_channels_max : 14320 + +enable_debug : false + +pe_array { + num_interleaved_features : 4 + num_interleaved_filters : 1 + exit_fifo_depth : 1024 +} + +filter_scratchpad { + filter_depth : 512 + bias_scale_depth : 512 +} + +dma { + csr_addr_width : 11 + csr_data_bytes : 4 + ddr_addr_width : 32 + ddr_burst_width : 4 + ddr_data_bytes : 64 + ddr_read_id_width : 2 +} + +activation { + generic_aux_parameters { + k_vector : 16 + } + enable_relu : true + enable_leaky_relu : false + enable_sigmoid : false + enable_prelu : false +} + +pool { + generic_aux_parameters { + k_vector : 1 + } + max_window_height : 3 + max_window_width : 3 + max_stride_vertical : 4 + max_stride_horizontal : 4 +} + +xbar { + xbar_k_vector : 16 + max_input_interfaces : 4 + max_output_interfaces : 4 + xbar_ports { + xbar_aux_port { + name : 'activation' + input_connection : 'xbar_in_port' + } + xbar_aux_port { + name : 'pool' + input_connection : 'xbar_in_port' + input_connection : 'activation' + } + } + xbar_in_port { + external_connection : 'pe_array' + } + xbar_out_port { + external_connection : 'input_feeder' + external_connection : 'output_writer' + input_connection : 'xbar_in_port' + input_connection : 'pool' + input_connection : 'activation' + } +} + +config_network { + modules { + module { + name : 'filter_reader' + cross_clock : 1 + } + module { + name : 'input_feeder_mux' + } + module { + name : 'input_feeder_writer' + } + module { + name : 'input_feeder_in' + } + module { + name : 'input_feeder_reader' + } + module { + name : 'input_feeder_out' + } + module { + name : 'feature_writer' + cross_clock : 1 + } + module { + name : 'feature_reader' + cross_clock : 1 + } + module { + name : 'xbar' + } + module { + name : 'activation' + } + module { + name : 'pool' + } + } +} diff --git a/python/openvino/demo/models/architectures/AGX7_Small_NoSoftmax.arch b/python/openvino/demo/models/architectures/AGX7_Small_NoSoftmax.arch new file mode 100644 index 0000000..531bc91 --- /dev/null +++ b/python/openvino/demo/models/architectures/AGX7_Small_NoSoftmax.arch @@ -0,0 +1,118 @@ +family : 'AGX7' +k_vector : 16 +c_vector : 16 +arch_precision : FP11 + +stream_buffer_depth : 17600 + +output_channels_max : 14320 + +enable_debug : false + +pe_array { + num_interleaved_features : 5 + num_interleaved_filters : 1 + exit_fifo_depth : 1024 +} + +filter_scratchpad { + filter_depth : 512 + bias_scale_depth : 512 +} + +dma { + csr_addr_width : 11 + csr_data_bytes : 4 + ddr_addr_width : 32 + ddr_burst_width : 4 + ddr_data_bytes : 64 + ddr_read_id_width : 2 +} + +activation { + generic_aux_parameters { + k_vector : 16 + } + enable_relu : true + enable_leaky_relu : false + enable_sigmoid : false + enable_prelu : false +} + +pool { + generic_aux_parameters { + k_vector : 1 + } + max_window_height : 3 + max_window_width : 3 + max_stride_vertical : 4 + max_stride_horizontal : 4 +} + +xbar { + xbar_k_vector : 16 + max_input_interfaces : 4 + max_output_interfaces : 4 + xbar_ports { + xbar_aux_port { + name : 'activation' + input_connection : 'xbar_in_port' + } + xbar_aux_port { + name : 'pool' + input_connection : 'xbar_in_port' + input_connection : 'activation' + } + } + xbar_in_port { + external_connection : 'pe_array' + } + xbar_out_port { + external_connection : 'input_feeder' + external_connection : 'output_writer' + input_connection : 'xbar_in_port' + input_connection : 'pool' + input_connection : 'activation' + } +} + +config_network { + modules { + module { + name : 'filter_reader' + cross_clock : 1 + } + module { + name : 'input_feeder_mux' + } + module { + name : 'input_feeder_writer' + } + module { + name : 'input_feeder_in' + } + module { + name : 'input_feeder_reader' + } + module { + name : 'input_feeder_out' + } + module { + name : 'feature_writer' + cross_clock : 1 + } + module { + name : 'feature_reader' + cross_clock : 1 + } + module { + name : 'xbar' + } + module { + name : 'activation' + } + module { + name : 'pool' + } + } +} diff --git a/python/openvino/demo/models/architectures/agx7_area.arch b/python/openvino/demo/models/architectures/agx7_area.arch new file mode 100644 index 0000000..4b64070 --- /dev/null +++ b/python/openvino/demo/models/architectures/agx7_area.arch @@ -0,0 +1,111 @@ +k_vector: 32 +c_vector: 32 +family: "AGX7" +arch_precision: FP11 +filter_precision: FP11 +sb_precision: FP11 +output_channels_max: 14320 +enable_debug: false +stream_buffer_depth: 6148 +dma { + csr_addr_width: 11 + csr_data_bytes: 4 + ddr_addr_width: 32 + ddr_burst_width: 4 + ddr_data_bytes: 64 + ddr_read_id_width: 2 +} +config_network { + modules { + module { + name: "filter_reader" + cross_clock: 1 + } + module { + name: "input_feeder_mux" + } + module { + name: "input_feeder_writer" + } + module { + name: "input_feeder_in" + } + module { + name: "input_feeder_reader" + } + module { + name: "input_feeder_out" + } + module { + name: "feature_writer" + cross_clock: 1 + } + module { + name: "feature_reader" + cross_clock: 1 + } + module { + name: "xbar" + } + module { + name: "activation" + } + module { + name: "pool" + } + } +} +xbar { + xbar_k_vector: 4 + max_input_interfaces: 4 + max_output_interfaces: 4 + xbar_ports { + xbar_aux_port { + name: "activation" + input_connection: "xbar_in_port" + } + xbar_aux_port { + name: "pool" + input_connection: "xbar_in_port" + input_connection: "activation" + } + } + xbar_in_port { + external_connection: "pe_array" + } + xbar_out_port { + external_connection: "input_feeder" + external_connection: "output_writer" + input_connection: "xbar_in_port" + input_connection: "pool" + input_connection: "activation" + } +} +activation { + generic_aux_parameters { + k_vector: 4 + } + enable_relu: true + enable_leaky_relu: false + enable_prelu: false + enable_sigmoid: false +} +pool { + generic_aux_parameters { + k_vector: 4 + } + max_window_height: 3 + max_window_width: 3 + max_stride_vertical: 4 + max_stride_horizontal: 4 +} +pe_array { + dsp_limit: 8528 + num_interleaved_features: 5 + num_interleaved_filters: 1 + exit_fifo_depth: 1024 +} +filter_scratchpad { + filter_depth: 512 + bias_scale_depth: 512 +} diff --git a/python/openvino/demo/models/architectures/agx7_area.ptc b/python/openvino/demo/models/architectures/agx7_area.ptc new file mode 100644 index 0000000..feb1a77 --- /dev/null +++ b/python/openvino/demo/models/architectures/agx7_area.ptc @@ -0,0 +1,248 @@ +{ + "blocks": { + "apply_margin_group": [ + { + "ThermalApplyMargin": "0%" + } + ], + "auto_compute_tj_group": [ + { + "Main_AUTO_COMPUTE_TJ": "Detailed Thermal Model" + } + ], + "calc_mode_group": [ + { + "ThermalCalcMode": "Find cooling solution for maximum junction temperature limit" + } + ], + "clock_row_group": [ + { + "CLOCKUSERINPUT_CLOCK_FREQ": "1000", + "CLOCKUSERINPUT_DOMAIN": "", + "CLOCKUSERINPUT_FANOUT": "95581", + "CLOCKUSERINPUT_GLOBAL_EN": "100%", + "CLOCKUSERINPUT_LOCAL_EN": "100%", + "CLOCKUSERINPUT_MODULE": "", + "CLOCKUSERINPUT_USER_COMMENTS": "", + "CLOCKUSERINPUT_UTILIZATION_FACTOR": "2" + } + ], + "cpu_group": [ + { + "HPSUSERINPUT_CPU_APP": "Off", + "HPSUSERINPUT_CPU_FREQUENCY": "0", + "HPSUSERINPUT_CPU_NUM_A53_CORES": "0", + "HPSUSERINPUT_CPU_NUM_A55_CORES": "0", + "HPSUSERINPUT_CPU_NUM_A76_CORES": "0", + "HPSUSERINPUT_L3_CACHE_SIZE": "0", + "HPSUSERINPUT_UNUSED_CORES_TREATMENT": "Power Up Unused Cores", + "HPS_INSTANCE_PATH": "", + "HPS_MODULE": "", + "HPS_SWITCH": "OFF", + "HPS_VCCL_HPSVoltage": "N/A" + } + ], + "device_definition": [ + { + "Device": "AGFB027R24C" + } + ], + "dsp_row_group": [ + { + "DSPUSERINPUT_CLOCK_ENABLE": "100%", + "DSPUSERINPUT_CLOCK_FREQ": "1000", + "DSPUSERINPUT_COEFFICIENT": "No", + "DSPUSERINPUT_CONFIG": "Sum of 4 9X9", + "DSPUSERINPUT_INSTANCE_PATH": "", + "DSPUSERINPUT_MODULE": "", + "DSPUSERINPUT_NUM_INSTANCES": "288", + "DSPUSERINPUT_PREADDER": "No", + "DSPUSERINPUT_REGISTERED_STAGES": "4", + "DSPUSERINPUT_TOGGLE_RATE": "12.5%", + "DSPUSERINPUT_USER_COMMENTS": "" + } + ], + "family_definition": [ + { + "Family": "Agilex 7" + } + ], + "hbm_freq_group": [ + { + "HBMUSERINPUT_HBM0_FREQUENCY": "0", + "HBMUSERINPUT_HBM1_FREQUENCY": "0" + } + ], + "io_unused_banks_group": [ + { + "IOUSERINPUT_UNUSED_HVIO_BANKS": "N/A", + "IOUSERINPUT_UNUSED_IO_BANKS": "1.2V" + } + ], + "logic_row_group": [ + { + "LOGICUSERINPUT_CLOCK_FREQ": "1000", + "LOGICUSERINPUT_INSTANCE_PATH": "", + "LOGICUSERINPUT_MODULE": "", + "LOGICUSERINPUT_NUM_FF": "56621", + "LOGICUSERINPUT_NUM_HALF_ALM": "36199", + "LOGICUSERINPUT_ROUTING_FACTOR": "3", + "LOGICUSERINPUT_TOGGLE_RATE": "12.5%", + "LOGICUSERINPUT_USER_COMMENTS": "" + } + ], + "package_definition": [ + { + "Package": "R24C" + } + ], + "power_char_definition": [ + { + "PowerChar": "Maximum" + } + ], + "ram_row_group": [ + { + "RAMUSERINPUT_BLOCK_ID": "N/A", + "RAMUSERINPUT_DEPTH": "512", + "RAMUSERINPUT_INSTANCE_PATH": "", + "RAMUSERINPUT_MODE": "Simple Dual Port", + "RAMUSERINPUT_MODULE": "", + "RAMUSERINPUT_NUM_BLOCKS": "232", + "RAMUSERINPUT_PORT_A_CLOCK_FREQ": "1000", + "RAMUSERINPUT_PORT_A_ENABLE": "100%", + "RAMUSERINPUT_PORT_A_RDEN": "0%", + "RAMUSERINPUT_PORT_A_WREN": "50%", + "RAMUSERINPUT_PORT_B_CLOCK_FREQ": "1000", + "RAMUSERINPUT_PORT_B_ENABLE": "100%", + "RAMUSERINPUT_PORT_B_RDEN": "100%", + "RAMUSERINPUT_PORT_B_WREN": "0%", + "RAMUSERINPUT_PORT_C_WREN": "0%", + "RAMUSERINPUT_PORT_D_RDEN": "0%", + "RAMUSERINPUT_TOGGLE_RATE": "12.5%", + "RAMUSERINPUT_TYPE": "M20K", + "RAMUSERINPUT_UNOC_MODE": "N/A", + "RAMUSERINPUT_UNOC_PORT_ENABLE": "N/A", + "RAMUSERINPUT_UNOC_PORT_WREN": "N/A", + "RAMUSERINPUT_USER_COMMENTS": "", + "RAMUSERINPUT_WIDTH": "40" + }, + { + "RAMUSERINPUT_BLOCK_ID": "N/A", + "RAMUSERINPUT_DEPTH": "1024", + "RAMUSERINPUT_INSTANCE_PATH": "", + "RAMUSERINPUT_MODE": "Simple Dual Port", + "RAMUSERINPUT_MODULE": "", + "RAMUSERINPUT_NUM_BLOCKS": "8", + "RAMUSERINPUT_PORT_A_CLOCK_FREQ": "1000", + "RAMUSERINPUT_PORT_A_ENABLE": "100%", + "RAMUSERINPUT_PORT_A_RDEN": "0%", + "RAMUSERINPUT_PORT_A_WREN": "50%", + "RAMUSERINPUT_PORT_B_CLOCK_FREQ": "1000", + "RAMUSERINPUT_PORT_B_ENABLE": "100%", + "RAMUSERINPUT_PORT_B_RDEN": "100%", + "RAMUSERINPUT_PORT_B_WREN": "0%", + "RAMUSERINPUT_PORT_C_WREN": "0%", + "RAMUSERINPUT_PORT_D_RDEN": "0%", + "RAMUSERINPUT_TOGGLE_RATE": "12.5%", + "RAMUSERINPUT_TYPE": "M20K", + "RAMUSERINPUT_UNOC_MODE": "N/A", + "RAMUSERINPUT_UNOC_PORT_ENABLE": "N/A", + "RAMUSERINPUT_UNOC_PORT_WREN": "N/A", + "RAMUSERINPUT_USER_COMMENTS": "", + "RAMUSERINPUT_WIDTH": "20" + }, + { + "RAMUSERINPUT_BLOCK_ID": "N/A", + "RAMUSERINPUT_DEPTH": "2048", + "RAMUSERINPUT_INSTANCE_PATH": "", + "RAMUSERINPUT_MODE": "Simple Dual Port", + "RAMUSERINPUT_MODULE": "", + "RAMUSERINPUT_NUM_BLOCKS": "69", + "RAMUSERINPUT_PORT_A_CLOCK_FREQ": "1000", + "RAMUSERINPUT_PORT_A_ENABLE": "100%", + "RAMUSERINPUT_PORT_A_RDEN": "0%", + "RAMUSERINPUT_PORT_A_WREN": "50%", + "RAMUSERINPUT_PORT_B_CLOCK_FREQ": "1000", + "RAMUSERINPUT_PORT_B_ENABLE": "100%", + "RAMUSERINPUT_PORT_B_RDEN": "100%", + "RAMUSERINPUT_PORT_B_WREN": "0%", + "RAMUSERINPUT_PORT_C_WREN": "0%", + "RAMUSERINPUT_PORT_D_RDEN": "0%", + "RAMUSERINPUT_TOGGLE_RATE": "12.5%", + "RAMUSERINPUT_TYPE": "M20K", + "RAMUSERINPUT_UNOC_MODE": "N/A", + "RAMUSERINPUT_UNOC_PORT_ENABLE": "N/A", + "RAMUSERINPUT_UNOC_PORT_WREN": "N/A", + "RAMUSERINPUT_USER_COMMENTS": "", + "RAMUSERINPUT_WIDTH": "10" + }, + { + "RAMUSERINPUT_BLOCK_ID": "N/A", + "RAMUSERINPUT_DEPTH": "2048", + "RAMUSERINPUT_INSTANCE_PATH": "", + "RAMUSERINPUT_MODE": "Simple Dual Port", + "RAMUSERINPUT_MODULE": "", + "RAMUSERINPUT_NUM_BLOCKS": "79", + "RAMUSERINPUT_PORT_A_CLOCK_FREQ": "1000", + "RAMUSERINPUT_PORT_A_ENABLE": "100%", + "RAMUSERINPUT_PORT_A_RDEN": "0%", + "RAMUSERINPUT_PORT_A_WREN": "50%", + "RAMUSERINPUT_PORT_B_CLOCK_FREQ": "1000", + "RAMUSERINPUT_PORT_B_ENABLE": "100%", + "RAMUSERINPUT_PORT_B_RDEN": "100%", + "RAMUSERINPUT_PORT_B_WREN": "0%", + "RAMUSERINPUT_PORT_C_WREN": "0%", + "RAMUSERINPUT_PORT_D_RDEN": "0%", + "RAMUSERINPUT_TOGGLE_RATE": "12.5%", + "RAMUSERINPUT_TYPE": "M20K", + "RAMUSERINPUT_UNOC_MODE": "N/A", + "RAMUSERINPUT_UNOC_PORT_ENABLE": "N/A", + "RAMUSERINPUT_UNOC_PORT_WREN": "N/A", + "RAMUSERINPUT_USER_COMMENTS": "", + "RAMUSERINPUT_WIDTH": "10" + } + ], + "temp_grade_definition": [ + { + "TempGrade": "Extended -2 Standard Power" + } + ], + "thermal_inputs_group": [ + { + "ThermalPsiCAInput": "N/A", + "ThermalTA": "25", + "ThermalTJMAX": "100" + } + ], + "thermal_parameters": [ + { + "Main_User_Tj": "N/A" + } + ], + "thermal_recalc_group": [ + { + "ThermalRecalcMode": "Automatic" + } + ], + "tsd_method_group": [ + { + "ThermalTsdMethod": "Not supported" + } + ], + "xcvr_grade_definition": [ + { + "XCVRGrade": "F2" + } + ], + "xcvr_unused_banks_group": [ + { + "HSSIUSERINPUT_UNUSED_HSSI_BANKS": "Power Down Unused Transceivers" + } + ] + }, + "header": { + "EPE_IMPORT_VERSION": "4.11" + }, + "ips": {} +} diff --git a/python/openvino/demo/models/architectures/agx7_performance.arch b/python/openvino/demo/models/architectures/agx7_performance.arch new file mode 100644 index 0000000..15e535a --- /dev/null +++ b/python/openvino/demo/models/architectures/agx7_performance.arch @@ -0,0 +1,111 @@ +k_vector: 128 +c_vector: 64 +family: "AGX7" +arch_precision: FP11 +filter_precision: FP11 +sb_precision: FP11 +output_channels_max: 14320 +enable_debug: false +stream_buffer_depth: 98441 +dma { + csr_addr_width: 11 + csr_data_bytes: 4 + ddr_addr_width: 32 + ddr_burst_width: 4 + ddr_data_bytes: 64 + ddr_read_id_width: 2 +} +config_network { + modules { + module { + name: "filter_reader" + cross_clock: 1 + } + module { + name: "input_feeder_mux" + } + module { + name: "input_feeder_writer" + } + module { + name: "input_feeder_in" + } + module { + name: "input_feeder_reader" + } + module { + name: "input_feeder_out" + } + module { + name: "feature_writer" + cross_clock: 1 + } + module { + name: "feature_reader" + cross_clock: 1 + } + module { + name: "xbar" + } + module { + name: "activation" + } + module { + name: "pool" + } + } +} +xbar { + xbar_k_vector: 64 + max_input_interfaces: 4 + max_output_interfaces: 4 + xbar_ports { + xbar_aux_port { + name: "activation" + input_connection: "xbar_in_port" + } + xbar_aux_port { + name: "pool" + input_connection: "xbar_in_port" + input_connection: "activation" + } + } + xbar_in_port { + external_connection: "pe_array" + } + xbar_out_port { + external_connection: "input_feeder" + external_connection: "output_writer" + input_connection: "xbar_in_port" + input_connection: "pool" + input_connection: "activation" + } +} +activation { + generic_aux_parameters { + k_vector: 64 + } + enable_relu: true + enable_leaky_relu: false + enable_prelu: false + enable_sigmoid: false +} +pool { + generic_aux_parameters { + k_vector: 64 + } + max_window_height: 3 + max_window_width: 3 + max_stride_vertical: 4 + max_stride_horizontal: 4 +} +pe_array { + dsp_limit: 8528 + num_interleaved_features: 5 + num_interleaved_filters: 1 + exit_fifo_depth: 1024 +} +filter_scratchpad { + filter_depth: 512 + bias_scale_depth: 512 +} diff --git a/python/openvino/demo/models/architectures/agx7_performance.ptc b/python/openvino/demo/models/architectures/agx7_performance.ptc new file mode 100644 index 0000000..00aad89 --- /dev/null +++ b/python/openvino/demo/models/architectures/agx7_performance.ptc @@ -0,0 +1,223 @@ +{ + "blocks": { + "apply_margin_group": [ + { + "ThermalApplyMargin": "0%" + } + ], + "auto_compute_tj_group": [ + { + "Main_AUTO_COMPUTE_TJ": "Detailed Thermal Model" + } + ], + "calc_mode_group": [ + { + "ThermalCalcMode": "Find cooling solution for maximum junction temperature limit" + } + ], + "clock_row_group": [ + { + "CLOCKUSERINPUT_CLOCK_FREQ": "1000", + "CLOCKUSERINPUT_DOMAIN": "", + "CLOCKUSERINPUT_FANOUT": "771050", + "CLOCKUSERINPUT_GLOBAL_EN": "100%", + "CLOCKUSERINPUT_LOCAL_EN": "100%", + "CLOCKUSERINPUT_MODULE": "", + "CLOCKUSERINPUT_USER_COMMENTS": "", + "CLOCKUSERINPUT_UTILIZATION_FACTOR": "2" + } + ], + "cpu_group": [ + { + "HPSUSERINPUT_CPU_APP": "Off", + "HPSUSERINPUT_CPU_FREQUENCY": "0", + "HPSUSERINPUT_CPU_NUM_A53_CORES": "0", + "HPSUSERINPUT_CPU_NUM_A55_CORES": "0", + "HPSUSERINPUT_CPU_NUM_A76_CORES": "0", + "HPSUSERINPUT_L3_CACHE_SIZE": "0", + "HPSUSERINPUT_UNUSED_CORES_TREATMENT": "Power Up Unused Cores", + "HPS_INSTANCE_PATH": "", + "HPS_MODULE": "", + "HPS_SWITCH": "OFF", + "HPS_VCCL_HPSVoltage": "N/A" + } + ], + "device_definition": [ + { + "Device": "AGFB027R24C" + } + ], + "dsp_row_group": [ + { + "DSPUSERINPUT_CLOCK_ENABLE": "100%", + "DSPUSERINPUT_CLOCK_FREQ": "1000", + "DSPUSERINPUT_COEFFICIENT": "No", + "DSPUSERINPUT_CONFIG": "Sum of 4 9X9", + "DSPUSERINPUT_INSTANCE_PATH": "", + "DSPUSERINPUT_MODULE": "", + "DSPUSERINPUT_NUM_INSTANCES": "2176", + "DSPUSERINPUT_PREADDER": "No", + "DSPUSERINPUT_REGISTERED_STAGES": "4", + "DSPUSERINPUT_TOGGLE_RATE": "12.5%", + "DSPUSERINPUT_USER_COMMENTS": "" + } + ], + "family_definition": [ + { + "Family": "Agilex 7" + } + ], + "hbm_freq_group": [ + { + "HBMUSERINPUT_HBM0_FREQUENCY": "0", + "HBMUSERINPUT_HBM1_FREQUENCY": "0" + } + ], + "io_unused_banks_group": [ + { + "IOUSERINPUT_UNUSED_HVIO_BANKS": "N/A", + "IOUSERINPUT_UNUSED_IO_BANKS": "1.2V" + } + ], + "logic_row_group": [ + { + "LOGICUSERINPUT_CLOCK_FREQ": "1000", + "LOGICUSERINPUT_INSTANCE_PATH": "", + "LOGICUSERINPUT_MODULE": "", + "LOGICUSERINPUT_NUM_FF": "551076", + "LOGICUSERINPUT_NUM_HALF_ALM": "217620", + "LOGICUSERINPUT_ROUTING_FACTOR": "3", + "LOGICUSERINPUT_TOGGLE_RATE": "12.5%", + "LOGICUSERINPUT_USER_COMMENTS": "" + } + ], + "package_definition": [ + { + "Package": "R24C" + } + ], + "power_char_definition": [ + { + "PowerChar": "Maximum" + } + ], + "ram_row_group": [ + { + "RAMUSERINPUT_BLOCK_ID": "N/A", + "RAMUSERINPUT_DEPTH": "512", + "RAMUSERINPUT_INSTANCE_PATH": "", + "RAMUSERINPUT_MODE": "Simple Dual Port", + "RAMUSERINPUT_MODULE": "", + "RAMUSERINPUT_NUM_BLOCKS": "1754", + "RAMUSERINPUT_PORT_A_CLOCK_FREQ": "1000", + "RAMUSERINPUT_PORT_A_ENABLE": "100%", + "RAMUSERINPUT_PORT_A_RDEN": "0%", + "RAMUSERINPUT_PORT_A_WREN": "50%", + "RAMUSERINPUT_PORT_B_CLOCK_FREQ": "1000", + "RAMUSERINPUT_PORT_B_ENABLE": "100%", + "RAMUSERINPUT_PORT_B_RDEN": "100%", + "RAMUSERINPUT_PORT_B_WREN": "0%", + "RAMUSERINPUT_PORT_C_WREN": "0%", + "RAMUSERINPUT_PORT_D_RDEN": "0%", + "RAMUSERINPUT_TOGGLE_RATE": "12.5%", + "RAMUSERINPUT_TYPE": "M20K", + "RAMUSERINPUT_UNOC_MODE": "N/A", + "RAMUSERINPUT_UNOC_PORT_ENABLE": "N/A", + "RAMUSERINPUT_UNOC_PORT_WREN": "N/A", + "RAMUSERINPUT_USER_COMMENTS": "", + "RAMUSERINPUT_WIDTH": "40" + }, + { + "RAMUSERINPUT_BLOCK_ID": "N/A", + "RAMUSERINPUT_DEPTH": "2048", + "RAMUSERINPUT_INSTANCE_PATH": "", + "RAMUSERINPUT_MODE": "Simple Dual Port", + "RAMUSERINPUT_MODULE": "", + "RAMUSERINPUT_NUM_BLOCKS": "2208", + "RAMUSERINPUT_PORT_A_CLOCK_FREQ": "1000", + "RAMUSERINPUT_PORT_A_ENABLE": "100%", + "RAMUSERINPUT_PORT_A_RDEN": "0%", + "RAMUSERINPUT_PORT_A_WREN": "50%", + "RAMUSERINPUT_PORT_B_CLOCK_FREQ": "1000", + "RAMUSERINPUT_PORT_B_ENABLE": "100%", + "RAMUSERINPUT_PORT_B_RDEN": "100%", + "RAMUSERINPUT_PORT_B_WREN": "0%", + "RAMUSERINPUT_PORT_C_WREN": "0%", + "RAMUSERINPUT_PORT_D_RDEN": "0%", + "RAMUSERINPUT_TOGGLE_RATE": "12.5%", + "RAMUSERINPUT_TYPE": "M20K", + "RAMUSERINPUT_UNOC_MODE": "N/A", + "RAMUSERINPUT_UNOC_PORT_ENABLE": "N/A", + "RAMUSERINPUT_UNOC_PORT_WREN": "N/A", + "RAMUSERINPUT_USER_COMMENTS": "", + "RAMUSERINPUT_WIDTH": "10" + }, + { + "RAMUSERINPUT_BLOCK_ID": "N/A", + "RAMUSERINPUT_DEPTH": "2048", + "RAMUSERINPUT_INSTANCE_PATH": "", + "RAMUSERINPUT_MODE": "Simple Dual Port", + "RAMUSERINPUT_MODULE": "", + "RAMUSERINPUT_NUM_BLOCKS": "129", + "RAMUSERINPUT_PORT_A_CLOCK_FREQ": "1000", + "RAMUSERINPUT_PORT_A_ENABLE": "100%", + "RAMUSERINPUT_PORT_A_RDEN": "0%", + "RAMUSERINPUT_PORT_A_WREN": "50%", + "RAMUSERINPUT_PORT_B_CLOCK_FREQ": "1000", + "RAMUSERINPUT_PORT_B_ENABLE": "100%", + "RAMUSERINPUT_PORT_B_RDEN": "100%", + "RAMUSERINPUT_PORT_B_WREN": "0%", + "RAMUSERINPUT_PORT_C_WREN": "0%", + "RAMUSERINPUT_PORT_D_RDEN": "0%", + "RAMUSERINPUT_TOGGLE_RATE": "12.5%", + "RAMUSERINPUT_TYPE": "M20K", + "RAMUSERINPUT_UNOC_MODE": "N/A", + "RAMUSERINPUT_UNOC_PORT_ENABLE": "N/A", + "RAMUSERINPUT_UNOC_PORT_WREN": "N/A", + "RAMUSERINPUT_USER_COMMENTS": "", + "RAMUSERINPUT_WIDTH": "10" + } + ], + "temp_grade_definition": [ + { + "TempGrade": "Extended -4 Extreme Low Power" + } + ], + "thermal_inputs_group": [ + { + "ThermalPsiCAInput": "N/A", + "ThermalTA": "25", + "ThermalTJMAX": "100" + } + ], + "thermal_parameters": [ + { + "Main_User_Tj": "N/A" + } + ], + "thermal_recalc_group": [ + { + "ThermalRecalcMode": "Automatic" + } + ], + "tsd_method_group": [ + { + "ThermalTsdMethod": "Not supported" + } + ], + "xcvr_grade_definition": [ + { + "XCVRGrade": "F2" + } + ], + "xcvr_unused_banks_group": [ + { + "HSSIUSERINPUT_UNUSED_HSSI_BANKS": "Power Down Unused Transceivers" + } + ] + }, + "header": { + "EPE_IMPORT_VERSION": "4.11" + }, + "ips": {} +} diff --git a/python/openvino/demo/models/architectures/de5a_area.arch b/python/openvino/demo/models/architectures/de5a_area.arch new file mode 100644 index 0000000..b881516 --- /dev/null +++ b/python/openvino/demo/models/architectures/de5a_area.arch @@ -0,0 +1,111 @@ +k_vector: 16 +c_vector: 4 +family: "A10" +arch_precision: FP11 +filter_precision: FP11 +sb_precision: FP11 +output_channels_max: 14320 +enable_debug: false +stream_buffer_depth: 5120 +dma { + csr_addr_width: 11 + csr_data_bytes: 4 + ddr_addr_width: 32 + ddr_burst_width: 4 + ddr_data_bytes: 64 + ddr_read_id_width: 2 +} +config_network { + modules { + module { + name: "filter_reader" + cross_clock: 1 + } + module { + name: "input_feeder_mux" + } + module { + name: "input_feeder_writer" + } + module { + name: "input_feeder_in" + } + module { + name: "input_feeder_reader" + } + module { + name: "input_feeder_out" + } + module { + name: "feature_writer" + cross_clock: 1 + } + module { + name: "feature_reader" + cross_clock: 1 + } + module { + name: "xbar" + } + module { + name: "activation" + } + module { + name: "pool" + } + } +} +xbar { + xbar_k_vector: 4 + max_input_interfaces: 4 + max_output_interfaces: 4 + xbar_ports { + xbar_aux_port { + name: "activation" + input_connection: "xbar_in_port" + } + xbar_aux_port { + name: "pool" + input_connection: "xbar_in_port" + input_connection: "activation" + } + } + xbar_in_port { + external_connection: "pe_array" + } + xbar_out_port { + external_connection: "input_feeder" + external_connection: "output_writer" + input_connection: "xbar_in_port" + input_connection: "pool" + input_connection: "activation" + } +} +activation { + generic_aux_parameters { + k_vector: 4 + } + enable_relu: true + enable_leaky_relu: false + enable_prelu: false + enable_sigmoid: false +} +pool { + generic_aux_parameters { + k_vector: 1 + } + max_window_height: 3 + max_window_width: 3 + max_stride_vertical: 4 + max_stride_horizontal: 4 +} +pe_array { + dsp_limit: 1518 + num_interleaved_features: 4 + num_interleaved_filters: 1 + exit_fifo_depth: 1024 +} +filter_scratchpad { + filter_depth: 512 + bias_scale_depth: 512 +} diff --git a/python/openvino/demo/models/architectures/de5a_performance.arch b/python/openvino/demo/models/architectures/de5a_performance.arch new file mode 100644 index 0000000..548e6be --- /dev/null +++ b/python/openvino/demo/models/architectures/de5a_performance.arch @@ -0,0 +1,111 @@ +k_vector: 64 +c_vector: 64 +family: "A10" +arch_precision: FP11 +filter_precision: FP11 +sb_precision: FP11 +output_channels_max: 14320 +enable_debug: false +stream_buffer_depth: 2088 +dma { + csr_addr_width: 11 + csr_data_bytes: 4 + ddr_addr_width: 32 + ddr_burst_width: 4 + ddr_data_bytes: 64 + ddr_read_id_width: 2 +} +config_network { + modules { + module { + name: "filter_reader" + cross_clock: 1 + } + module { + name: "input_feeder_mux" + } + module { + name: "input_feeder_writer" + } + module { + name: "input_feeder_in" + } + module { + name: "input_feeder_reader" + } + module { + name: "input_feeder_out" + } + module { + name: "feature_writer" + cross_clock: 1 + } + module { + name: "feature_reader" + cross_clock: 1 + } + module { + name: "xbar" + } + module { + name: "activation" + } + module { + name: "pool" + } + } +} +xbar { + xbar_k_vector: 64 + max_input_interfaces: 4 + max_output_interfaces: 4 + xbar_ports { + xbar_aux_port { + name: "activation" + input_connection: "xbar_in_port" + } + xbar_aux_port { + name: "pool" + input_connection: "xbar_in_port" + input_connection: "activation" + } + } + xbar_in_port { + external_connection: "pe_array" + } + xbar_out_port { + external_connection: "input_feeder" + external_connection: "output_writer" + input_connection: "xbar_in_port" + input_connection: "pool" + input_connection: "activation" + } +} +activation { + generic_aux_parameters { + k_vector: 64 + } + enable_relu: true + enable_leaky_relu: false + enable_prelu: false + enable_sigmoid: false +} +pool { + generic_aux_parameters { + k_vector: 64 + } + max_window_height: 3 + max_window_width: 3 + max_stride_vertical: 4 + max_stride_horizontal: 4 +} +pe_array { + dsp_limit: 1518 + num_interleaved_features: 4 + num_interleaved_filters: 1 + exit_fifo_depth: 1024 +} +filter_scratchpad { + filter_depth: 512 + bias_scale_depth: 512 +} diff --git a/python/openvino/demo/models/architectures/quartus_ptc.log b/python/openvino/demo/models/architectures/quartus_ptc.log new file mode 100644 index 0000000..ee09f4b --- /dev/null +++ b/python/openvino/demo/models/architectures/quartus_ptc.log @@ -0,0 +1,26 @@ +# ************************************************** +# Quartus Prime Version 24.3.0 +# Process ID 466429 +# Start logging UI actions +# Fri Jan 24 10:32:12 2025 +# ************************************************** +[Fri Jan 24 10:32:16 2025] Window title: 'Device selection' | 'majorFilter_0' combo box changed to: 'Agilex 7' +[Fri Jan 24 10:32:38 2025] Window title: 'Device selection' | QPushButton '&Close' pressed! +[Fri Jan 24 10:32:40 2025] Window title: 'PTC Error' | QPushButton '&OK' pressed! +[Fri Jan 24 10:32:55 2025] Window title: 'PTC Error' | QPushButton '&OK' pressed! +[Fri Jan 24 10:32:56 2025] Window title: 'Quartus Prime Power and Thermal Calculator Pro Edition - [agx7_area.ptc][*]' | 'Device' combo box changed to: 'AGFB014R24A' +[Fri Jan 24 10:35:19 2025] Window title: 'Device selection' | 'majorFilter_0' combo box changed to: 'Agilex 7' +[Fri Jan 24 10:36:24 2025] Window title: 'Device selection' | QPushButton 'Select Device' pressed! +[Fri Jan 24 10:37:10 2025] Window title: 'PTC Error' | QPushButton '&OK' pressed! +[Fri Jan 24 10:37:13 2025] Window title: 'Quartus Prime Power and Thermal Calculator Pro Edition - [agx7_area.ptc][*]' | 'ThermalRecalcMode' combo box changed to: 'Automatic' +[Fri Jan 24 10:37:15 2025] Window title: 'Quartus Prime Power and Thermal Calculator Pro Edition - [agx7_area.ptc][*]' | 'ThermalRecalcMode' combo box changed to: 'Manual' +[Fri Jan 24 10:40:30 2025] Window title: 'PTC Error' | QPushButton '&OK' pressed! +[Fri Jan 24 10:40:36 2025] [Action triggered] Window title: 'Quartus Prime Power and Thermal Calculator Pro Edition - [agx7_area.ptc][*]' | &File > '&Open...' was pressed +[Fri Jan 24 10:40:38 2025] Window title: 'Open PTC File' | QPushButton '&Open' pressed! +[Fri Jan 24 10:40:40 2025] Window title: 'Quartus Prime' | QPushButton '&No' pressed! +[Fri Jan 24 10:40:45 2025] Window title: 'PTC Error' | QPushButton '&OK' pressed! +[Fri Jan 24 10:40:52 2025] Window title: 'Quartus Prime' | QPushButton '&No' pressed! +# ************************************************** +# Fri Jan 24 10:40:52 2025 +# End of logging UI actions +# ************************************************** diff --git a/python/openvino/demo/models/architectures/quartus_ptc.rec b/python/openvino/demo/models/architectures/quartus_ptc.rec new file mode 100644 index 0000000..9221267 --- /dev/null +++ b/python/openvino/demo/models/architectures/quartus_ptc.rec @@ -0,0 +1,10 @@ +# ************************************************** +# Quartus Prime Version 24.3.0 +# Process ID 466429 +# Start logging TCL commands +# Fri Jan 24 10:32:12 2025 +# ************************************************** +# ************************************************** +# Fri Jan 24 10:40:52 2025 +# End of logging TCL commands +# ************************************************** diff --git a/python/openvino/demo/models/model_bmp.bin b/python/openvino/demo/models/model_bmp.bin Binary files differnew file mode 100644 index 0000000..12bf262 --- /dev/null +++ b/python/openvino/demo/models/model_bmp.bin diff --git a/python/openvino/demo/models/model_bmp.xml b/python/openvino/demo/models/model_bmp.xml new file mode 100644 index 0000000..7cd188d --- /dev/null +++ b/python/openvino/demo/models/model_bmp.xml @@ -0,0 +1,1782 @@ +<?xml version="1.0"?> +<net name="main_graph" version="11"> + <layers> + <layer id="0" name="input.1" type="Parameter" version="opset1"> + <data shape="32,3,128,128" element_type="f32" /> + <output> + <port id="0" precision="FP32" names="input.1"> + <dim>32</dim> + <dim>3</dim> + <dim>128</dim> + <dim>128</dim> + </port> + </output> + </layer> + <layer id="1" name="onnx::Conv_174" type="Const" version="opset1"> + <data element_type="f32" shape="16, 3, 3, 3" offset="0" size="1728" /> + <output> + <port id="0" precision="FP32" names="onnx::Conv_174"> + <dim>16</dim> + <dim>3</dim> + <dim>3</dim> + <dim>3</dim> + </port> + </output> + </layer> + <layer id="2" name="/conv2x_0/conv2x_0.0/Conv/WithoutBiases" type="Convolution" version="opset1"> + <data strides="1, 1" dilations="1, 1" pads_begin="1, 1" pads_end="1, 1" auto_pad="explicit" /> + <input> + <port id="0" precision="FP32"> + <dim>32</dim> + <dim>3</dim> + <dim>128</dim> + <dim>128</dim> + </port> + <port id="1" precision="FP32"> + <dim>16</dim> + <dim>3</dim> + <dim>3</dim> + <dim>3</dim> + </port> + </input> + <output> + <port id="2" precision="FP32"> + <dim>32</dim> + <dim>16</dim> + <dim>128</dim> + <dim>128</dim> + </port> + </output> + </layer> + <layer id="3" name="Reshape_44" type="Const" version="opset1"> + <data element_type="f32" shape="1, 16, 1, 1" offset="1728" size="64" /> + <output> + <port id="0" precision="FP32"> + <dim>1</dim> + <dim>16</dim> + <dim>1</dim> + <dim>1</dim> + </port> + </output> + </layer> + <layer id="4" name="/conv2x_0/conv2x_0.0/Conv" type="Add" version="opset1"> + <data auto_broadcast="numpy" /> + <input> + <port id="0" precision="FP32"> + <dim>32</dim> + <dim>16</dim> + <dim>128</dim> + <dim>128</dim> + </port> + <port id="1" precision="FP32"> + <dim>1</dim> + <dim>16</dim> + <dim>1</dim> + <dim>1</dim> + </port> + </input> + <output> + <port id="2" precision="FP32" names="/conv2x_0/conv2x_0.0/Conv_output_0"> + <dim>32</dim> + <dim>16</dim> + <dim>128</dim> + <dim>128</dim> + </port> + </output> + </layer> + <layer id="5" name="/conv2x_0/conv2x_0.2/Relu" type="ReLU" version="opset1"> + <input> + <port id="0" precision="FP32"> + <dim>32</dim> + <dim>16</dim> + <dim>128</dim> + <dim>128</dim> + </port> + </input> + <output> + <port id="1" precision="FP32" names="/conv2x_0/conv2x_0.2/Relu_output_0"> + <dim>32</dim> + <dim>16</dim> + <dim>128</dim> + <dim>128</dim> + </port> + </output> + </layer> + <layer id="6" name="onnx::Conv_177" type="Const" version="opset1"> + <data element_type="f32" shape="16, 16, 3, 3" offset="1792" size="9216" /> + <output> + <port id="0" precision="FP32" names="onnx::Conv_177"> + <dim>16</dim> + <dim>16</dim> + <dim>3</dim> + <dim>3</dim> + </port> + </output> + </layer> + <layer id="7" name="/conv2x_0/conv2x_0.3/Conv/WithoutBiases" type="Convolution" version="opset1"> + <data strides="1, 1" dilations="1, 1" pads_begin="1, 1" pads_end="1, 1" auto_pad="explicit" /> + <input> + <port id="0" precision="FP32"> + <dim>32</dim> + <dim>16</dim> + <dim>128</dim> + <dim>128</dim> + </port> + <port id="1" precision="FP32"> + <dim>16</dim> + <dim>16</dim> + <dim>3</dim> + <dim>3</dim> + </port> + </input> + <output> + <port id="2" precision="FP32"> + <dim>32</dim> + <dim>16</dim> + <dim>128</dim> + <dim>128</dim> + </port> + </output> + </layer> + <layer id="8" name="Reshape_61" type="Const" version="opset1"> + <data element_type="f32" shape="1, 16, 1, 1" offset="11008" size="64" /> + <output> + <port id="0" precision="FP32"> + <dim>1</dim> + <dim>16</dim> + <dim>1</dim> + <dim>1</dim> + </port> + </output> + </layer> + <layer id="9" name="/conv2x_0/conv2x_0.3/Conv" type="Add" version="opset1"> + <data auto_broadcast="numpy" /> + <input> + <port id="0" precision="FP32"> + <dim>32</dim> + <dim>16</dim> + <dim>128</dim> + <dim>128</dim> + </port> + <port id="1" precision="FP32"> + <dim>1</dim> + <dim>16</dim> + <dim>1</dim> + <dim>1</dim> + </port> + </input> + <output> + <port id="2" precision="FP32" names="/conv2x_0/conv2x_0.3/Conv_output_0"> + <dim>32</dim> + <dim>16</dim> + <dim>128</dim> + <dim>128</dim> + </port> + </output> + </layer> + <layer id="10" name="/conv2x_0/conv2x_0.5/Relu" type="ReLU" version="opset1"> + <input> + <port id="0" precision="FP32"> + <dim>32</dim> + <dim>16</dim> + <dim>128</dim> + <dim>128</dim> + </port> + </input> + <output> + <port id="1" precision="FP32" names="/conv2x_0/conv2x_0.5/Relu_output_0"> + <dim>32</dim> + <dim>16</dim> + <dim>128</dim> + <dim>128</dim> + </port> + </output> + </layer> + <layer id="11" name="/pool/MaxPool" type="MaxPool" version="opset8"> + <data strides="2, 2" dilations="1, 1" pads_begin="0, 0" pads_end="0, 0" kernel="2, 2" rounding_type="floor" auto_pad="explicit" index_element_type="i64" axis="0" /> + <input> + <port id="0" precision="FP32"> + <dim>32</dim> + <dim>16</dim> + <dim>128</dim> + <dim>128</dim> + </port> + </input> + <output> + <port id="1" precision="FP32" names="/pool/MaxPool_output_0"> + <dim>32</dim> + <dim>16</dim> + <dim>64</dim> + <dim>64</dim> + </port> + <port id="2" precision="I64"> + <dim>32</dim> + <dim>16</dim> + <dim>64</dim> + <dim>64</dim> + </port> + </output> + </layer> + <layer id="12" name="onnx::Conv_180" type="Const" version="opset1"> + <data element_type="f32" shape="32, 16, 3, 3" offset="11072" size="18432" /> + <output> + <port id="0" precision="FP32" names="onnx::Conv_180"> + <dim>32</dim> + <dim>16</dim> + <dim>3</dim> + <dim>3</dim> + </port> + </output> + </layer> + <layer id="13" name="/conv2x_1/conv2x_1.0/Conv/WithoutBiases" type="Convolution" version="opset1"> + <data strides="1, 1" dilations="1, 1" pads_begin="1, 1" pads_end="1, 1" auto_pad="explicit" /> + <input> + <port id="0" precision="FP32"> + <dim>32</dim> + <dim>16</dim> + <dim>64</dim> + <dim>64</dim> + </port> + <port id="1" precision="FP32"> + <dim>32</dim> + <dim>16</dim> + <dim>3</dim> + <dim>3</dim> + </port> + </input> + <output> + <port id="2" precision="FP32"> + <dim>32</dim> + <dim>32</dim> + <dim>64</dim> + <dim>64</dim> + </port> + </output> + </layer> + <layer id="14" name="Reshape_79" type="Const" version="opset1"> + <data element_type="f32" shape="1, 32, 1, 1" offset="29504" size="128" /> + <output> + <port id="0" precision="FP32"> + <dim>1</dim> + <dim>32</dim> + <dim>1</dim> + <dim>1</dim> + </port> + </output> + </layer> + <layer id="15" name="/conv2x_1/conv2x_1.0/Conv" type="Add" version="opset1"> + <data auto_broadcast="numpy" /> + <input> + <port id="0" precision="FP32"> + <dim>32</dim> + <dim>32</dim> + <dim>64</dim> + <dim>64</dim> + </port> + <port id="1" precision="FP32"> + <dim>1</dim> + <dim>32</dim> + <dim>1</dim> + <dim>1</dim> + </port> + </input> + <output> + <port id="2" precision="FP32" names="/conv2x_1/conv2x_1.0/Conv_output_0"> + <dim>32</dim> + <dim>32</dim> + <dim>64</dim> + <dim>64</dim> + </port> + </output> + </layer> + <layer id="16" name="/conv2x_1/conv2x_1.2/Relu" type="ReLU" version="opset1"> + <input> + <port id="0" precision="FP32"> + <dim>32</dim> + <dim>32</dim> + <dim>64</dim> + <dim>64</dim> + </port> + </input> + <output> + <port id="1" precision="FP32" names="/conv2x_1/conv2x_1.2/Relu_output_0"> + <dim>32</dim> + <dim>32</dim> + <dim>64</dim> + <dim>64</dim> + </port> + </output> + </layer> + <layer id="17" name="onnx::Conv_183" type="Const" version="opset1"> + <data element_type="f32" shape="32, 32, 3, 3" offset="29632" size="36864" /> + <output> + <port id="0" precision="FP32" names="onnx::Conv_183"> + <dim>32</dim> + <dim>32</dim> + <dim>3</dim> + <dim>3</dim> + </port> + </output> + </layer> + <layer id="18" name="/conv2x_1/conv2x_1.3/Conv/WithoutBiases" type="Convolution" version="opset1"> + <data strides="1, 1" dilations="1, 1" pads_begin="1, 1" pads_end="1, 1" auto_pad="explicit" /> + <input> + <port id="0" precision="FP32"> + <dim>32</dim> + <dim>32</dim> + <dim>64</dim> + <dim>64</dim> + </port> + <port id="1" precision="FP32"> + <dim>32</dim> + <dim>32</dim> + <dim>3</dim> + <dim>3</dim> + </port> + </input> + <output> + <port id="2" precision="FP32"> + <dim>32</dim> + <dim>32</dim> + <dim>64</dim> + <dim>64</dim> + </port> + </output> + </layer> + <layer id="19" name="Reshape_96" type="Const" version="opset1"> + <data element_type="f32" shape="1, 32, 1, 1" offset="66496" size="128" /> + <output> + <port id="0" precision="FP32"> + <dim>1</dim> + <dim>32</dim> + <dim>1</dim> + <dim>1</dim> + </port> + </output> + </layer> + <layer id="20" name="/conv2x_1/conv2x_1.3/Conv" type="Add" version="opset1"> + <data auto_broadcast="numpy" /> + <input> + <port id="0" precision="FP32"> + <dim>32</dim> + <dim>32</dim> + <dim>64</dim> + <dim>64</dim> + </port> + <port id="1" precision="FP32"> + <dim>1</dim> + <dim>32</dim> + <dim>1</dim> + <dim>1</dim> + </port> + </input> + <output> + <port id="2" precision="FP32" names="/conv2x_1/conv2x_1.3/Conv_output_0"> + <dim>32</dim> + <dim>32</dim> + <dim>64</dim> + <dim>64</dim> + </port> + </output> + </layer> + <layer id="21" name="/conv2x_1/conv2x_1.5/Relu" type="ReLU" version="opset1"> + <input> + <port id="0" precision="FP32"> + <dim>32</dim> + <dim>32</dim> + <dim>64</dim> + <dim>64</dim> + </port> + </input> + <output> + <port id="1" precision="FP32" names="/conv2x_1/conv2x_1.5/Relu_output_0"> + <dim>32</dim> + <dim>32</dim> + <dim>64</dim> + <dim>64</dim> + </port> + </output> + </layer> + <layer id="22" name="/pool_1/MaxPool" type="MaxPool" version="opset8"> + <data strides="2, 2" dilations="1, 1" pads_begin="0, 0" pads_end="0, 0" kernel="2, 2" rounding_type="floor" auto_pad="explicit" index_element_type="i64" axis="0" /> + <input> + <port id="0" precision="FP32"> + <dim>32</dim> + <dim>32</dim> + <dim>64</dim> + <dim>64</dim> + </port> + </input> + <output> + <port id="1" precision="FP32" names="/pool_1/MaxPool_output_0"> + <dim>32</dim> + <dim>32</dim> + <dim>32</dim> + <dim>32</dim> + </port> + <port id="2" precision="I64"> + <dim>32</dim> + <dim>32</dim> + <dim>32</dim> + <dim>32</dim> + </port> + </output> + </layer> + <layer id="23" name="onnx::Conv_186" type="Const" version="opset1"> + <data element_type="f32" shape="64, 32, 3, 3" offset="66624" size="73728" /> + <output> + <port id="0" precision="FP32" names="onnx::Conv_186"> + <dim>64</dim> + <dim>32</dim> + <dim>3</dim> + <dim>3</dim> + </port> + </output> + </layer> + <layer id="24" name="/conv2x_2/conv2x_2.0/Conv/WithoutBiases" type="Convolution" version="opset1"> + <data strides="1, 1" dilations="1, 1" pads_begin="1, 1" pads_end="1, 1" auto_pad="explicit" /> + <input> + <port id="0" precision="FP32"> + <dim>32</dim> + <dim>32</dim> + <dim>32</dim> + <dim>32</dim> + </port> + <port id="1" precision="FP32"> + <dim>64</dim> + <dim>32</dim> + <dim>3</dim> + <dim>3</dim> + </port> + </input> + <output> + <port id="2" precision="FP32"> + <dim>32</dim> + <dim>64</dim> + <dim>32</dim> + <dim>32</dim> + </port> + </output> + </layer> + <layer id="25" name="Reshape_114" type="Const" version="opset1"> + <data element_type="f32" shape="1, 64, 1, 1" offset="140352" size="256" /> + <output> + <port id="0" precision="FP32"> + <dim>1</dim> + <dim>64</dim> + <dim>1</dim> + <dim>1</dim> + </port> + </output> + </layer> + <layer id="26" name="/conv2x_2/conv2x_2.0/Conv" type="Add" version="opset1"> + <data auto_broadcast="numpy" /> + <input> + <port id="0" precision="FP32"> + <dim>32</dim> + <dim>64</dim> + <dim>32</dim> + <dim>32</dim> + </port> + <port id="1" precision="FP32"> + <dim>1</dim> + <dim>64</dim> + <dim>1</dim> + <dim>1</dim> + </port> + </input> + <output> + <port id="2" precision="FP32" names="/conv2x_2/conv2x_2.0/Conv_output_0"> + <dim>32</dim> + <dim>64</dim> + <dim>32</dim> + <dim>32</dim> + </port> + </output> + </layer> + <layer id="27" name="/conv2x_2/conv2x_2.2/Relu" type="ReLU" version="opset1"> + <input> + <port id="0" precision="FP32"> + <dim>32</dim> + <dim>64</dim> + <dim>32</dim> + <dim>32</dim> + </port> + </input> + <output> + <port id="1" precision="FP32" names="/conv2x_2/conv2x_2.2/Relu_output_0"> + <dim>32</dim> + <dim>64</dim> + <dim>32</dim> + <dim>32</dim> + </port> + </output> + </layer> + <layer id="28" name="onnx::Conv_189" type="Const" version="opset1"> + <data element_type="f32" shape="64, 64, 3, 3" offset="140608" size="147456" /> + <output> + <port id="0" precision="FP32" names="onnx::Conv_189"> + <dim>64</dim> + <dim>64</dim> + <dim>3</dim> + <dim>3</dim> + </port> + </output> + </layer> + <layer id="29" name="/conv2x_2/conv2x_2.3/Conv/WithoutBiases" type="Convolution" version="opset1"> + <data strides="1, 1" dilations="1, 1" pads_begin="1, 1" pads_end="1, 1" auto_pad="explicit" /> + <input> + <port id="0" precision="FP32"> + <dim>32</dim> + <dim>64</dim> + <dim>32</dim> + <dim>32</dim> + </port> + <port id="1" precision="FP32"> + <dim>64</dim> + <dim>64</dim> + <dim>3</dim> + <dim>3</dim> + </port> + </input> + <output> + <port id="2" precision="FP32"> + <dim>32</dim> + <dim>64</dim> + <dim>32</dim> + <dim>32</dim> + </port> + </output> + </layer> + <layer id="30" name="Reshape_131" type="Const" version="opset1"> + <data element_type="f32" shape="1, 64, 1, 1" offset="288064" size="256" /> + <output> + <port id="0" precision="FP32"> + <dim>1</dim> + <dim>64</dim> + <dim>1</dim> + <dim>1</dim> + </port> + </output> + </layer> + <layer id="31" name="/conv2x_2/conv2x_2.3/Conv" type="Add" version="opset1"> + <data auto_broadcast="numpy" /> + <input> + <port id="0" precision="FP32"> + <dim>32</dim> + <dim>64</dim> + <dim>32</dim> + <dim>32</dim> + </port> + <port id="1" precision="FP32"> + <dim>1</dim> + <dim>64</dim> + <dim>1</dim> + <dim>1</dim> + </port> + </input> + <output> + <port id="2" precision="FP32" names="/conv2x_2/conv2x_2.3/Conv_output_0"> + <dim>32</dim> + <dim>64</dim> + <dim>32</dim> + <dim>32</dim> + </port> + </output> + </layer> + <layer id="32" name="/conv2x_2/conv2x_2.5/Relu" type="ReLU" version="opset1"> + <input> + <port id="0" precision="FP32"> + <dim>32</dim> + <dim>64</dim> + <dim>32</dim> + <dim>32</dim> + </port> + </input> + <output> + <port id="1" precision="FP32" names="/conv2x_2/conv2x_2.5/Relu_output_0"> + <dim>32</dim> + <dim>64</dim> + <dim>32</dim> + <dim>32</dim> + </port> + </output> + </layer> + <layer id="33" name="/pool_2/MaxPool" type="MaxPool" version="opset8"> + <data strides="2, 2" dilations="1, 1" pads_begin="0, 0" pads_end="0, 0" kernel="2, 2" rounding_type="floor" auto_pad="explicit" index_element_type="i64" axis="0" /> + <input> + <port id="0" precision="FP32"> + <dim>32</dim> + <dim>64</dim> + <dim>32</dim> + <dim>32</dim> + </port> + </input> + <output> + <port id="1" precision="FP32" names="/pool_2/MaxPool_output_0"> + <dim>32</dim> + <dim>64</dim> + <dim>16</dim> + <dim>16</dim> + </port> + <port id="2" precision="I64"> + <dim>32</dim> + <dim>64</dim> + <dim>16</dim> + <dim>16</dim> + </port> + </output> + </layer> + <layer id="34" name="onnx::Conv_192" type="Const" version="opset1"> + <data element_type="f32" shape="128, 64, 3, 3" offset="288320" size="294912" /> + <output> + <port id="0" precision="FP32" names="onnx::Conv_192"> + <dim>128</dim> + <dim>64</dim> + <dim>3</dim> + <dim>3</dim> + </port> + </output> + </layer> + <layer id="35" name="/conv2x_3/conv2x_3.0/Conv/WithoutBiases" type="Convolution" version="opset1"> + <data strides="1, 1" dilations="1, 1" pads_begin="1, 1" pads_end="1, 1" auto_pad="explicit" /> + <input> + <port id="0" precision="FP32"> + <dim>32</dim> + <dim>64</dim> + <dim>16</dim> + <dim>16</dim> + </port> + <port id="1" precision="FP32"> + <dim>128</dim> + <dim>64</dim> + <dim>3</dim> + <dim>3</dim> + </port> + </input> + <output> + <port id="2" precision="FP32"> + <dim>32</dim> + <dim>128</dim> + <dim>16</dim> + <dim>16</dim> + </port> + </output> + </layer> + <layer id="36" name="Reshape_149" type="Const" version="opset1"> + <data element_type="f32" shape="1, 128, 1, 1" offset="583232" size="512" /> + <output> + <port id="0" precision="FP32"> + <dim>1</dim> + <dim>128</dim> + <dim>1</dim> + <dim>1</dim> + </port> + </output> + </layer> + <layer id="37" name="/conv2x_3/conv2x_3.0/Conv" type="Add" version="opset1"> + <data auto_broadcast="numpy" /> + <input> + <port id="0" precision="FP32"> + <dim>32</dim> + <dim>128</dim> + <dim>16</dim> + <dim>16</dim> + </port> + <port id="1" precision="FP32"> + <dim>1</dim> + <dim>128</dim> + <dim>1</dim> + <dim>1</dim> + </port> + </input> + <output> + <port id="2" precision="FP32" names="/conv2x_3/conv2x_3.0/Conv_output_0"> + <dim>32</dim> + <dim>128</dim> + <dim>16</dim> + <dim>16</dim> + </port> + </output> + </layer> + <layer id="38" name="/conv2x_3/conv2x_3.2/Relu" type="ReLU" version="opset1"> + <input> + <port id="0" precision="FP32"> + <dim>32</dim> + <dim>128</dim> + <dim>16</dim> + <dim>16</dim> + </port> + </input> + <output> + <port id="1" precision="FP32" names="/conv2x_3/conv2x_3.2/Relu_output_0"> + <dim>32</dim> + <dim>128</dim> + <dim>16</dim> + <dim>16</dim> + </port> + </output> + </layer> + <layer id="39" name="onnx::Conv_195" type="Const" version="opset1"> + <data element_type="f32" shape="128, 128, 3, 3" offset="583744" size="589824" /> + <output> + <port id="0" precision="FP32" names="onnx::Conv_195"> + <dim>128</dim> + <dim>128</dim> + <dim>3</dim> + <dim>3</dim> + </port> + </output> + </layer> + <layer id="40" name="/conv2x_3/conv2x_3.3/Conv/WithoutBiases" type="Convolution" version="opset1"> + <data strides="1, 1" dilations="1, 1" pads_begin="1, 1" pads_end="1, 1" auto_pad="explicit" /> + <input> + <port id="0" precision="FP32"> + <dim>32</dim> + <dim>128</dim> + <dim>16</dim> + <dim>16</dim> + </port> + <port id="1" precision="FP32"> + <dim>128</dim> + <dim>128</dim> + <dim>3</dim> + <dim>3</dim> + </port> + </input> + <output> + <port id="2" precision="FP32"> + <dim>32</dim> + <dim>128</dim> + <dim>16</dim> + <dim>16</dim> + </port> + </output> + </layer> + <layer id="41" name="Reshape_166" type="Const" version="opset1"> + <data element_type="f32" shape="1, 128, 1, 1" offset="1173568" size="512" /> + <output> + <port id="0" precision="FP32"> + <dim>1</dim> + <dim>128</dim> + <dim>1</dim> + <dim>1</dim> + </port> + </output> + </layer> + <layer id="42" name="/conv2x_3/conv2x_3.3/Conv" type="Add" version="opset1"> + <data auto_broadcast="numpy" /> + <input> + <port id="0" precision="FP32"> + <dim>32</dim> + <dim>128</dim> + <dim>16</dim> + <dim>16</dim> + </port> + <port id="1" precision="FP32"> + <dim>1</dim> + <dim>128</dim> + <dim>1</dim> + <dim>1</dim> + </port> + </input> + <output> + <port id="2" precision="FP32" names="/conv2x_3/conv2x_3.3/Conv_output_0"> + <dim>32</dim> + <dim>128</dim> + <dim>16</dim> + <dim>16</dim> + </port> + </output> + </layer> + <layer id="43" name="/conv2x_3/conv2x_3.5/Relu" type="ReLU" version="opset1"> + <input> + <port id="0" precision="FP32"> + <dim>32</dim> + <dim>128</dim> + <dim>16</dim> + <dim>16</dim> + </port> + </input> + <output> + <port id="1" precision="FP32" names="/conv2x_3/conv2x_3.5/Relu_output_0"> + <dim>32</dim> + <dim>128</dim> + <dim>16</dim> + <dim>16</dim> + </port> + </output> + </layer> + <layer id="44" name="/upsample/Constant" type="Const" version="opset1"> + <data element_type="f32" shape="4" offset="1174080" size="16" /> + <output> + <port id="0" precision="FP32" names="/upsample/Constant_output_0"> + <dim>4</dim> + </port> + </output> + </layer> + <layer id="45" name="/upsample/Resize" type="Interpolate" version="opset11"> + <data mode="nearest" shape_calculation_mode="scales" coordinate_transformation_mode="asymmetric" nearest_mode="floor" antialias="false" pads_begin="0, 0, 0, 0" pads_end="0, 0, 0, 0" cube_coeff="-0.75" /> + <input> + <port id="0" precision="FP32"> + <dim>32</dim> + <dim>128</dim> + <dim>16</dim> + <dim>16</dim> + </port> + <port id="1" precision="FP32"> + <dim>4</dim> + </port> + </input> + <output> + <port id="2" precision="FP32" names="/upsample/Resize_output_0"> + <dim>32</dim> + <dim>128</dim> + <dim>32</dim> + <dim>32</dim> + </port> + </output> + </layer> + <layer id="46" name="/Concat" type="Concat" version="opset1"> + <data axis="1" /> + <input> + <port id="0" precision="FP32"> + <dim>32</dim> + <dim>128</dim> + <dim>32</dim> + <dim>32</dim> + </port> + <port id="1" precision="FP32"> + <dim>32</dim> + <dim>64</dim> + <dim>32</dim> + <dim>32</dim> + </port> + </input> + <output> + <port id="2" precision="FP32" names="/Concat_output_0"> + <dim>32</dim> + <dim>192</dim> + <dim>32</dim> + <dim>32</dim> + </port> + </output> + </layer> + <layer id="47" name="onnx::Conv_198" type="Const" version="opset1"> + <data element_type="f32" shape="64, 192, 3, 3" offset="1174096" size="442368" /> + <output> + <port id="0" precision="FP32" names="onnx::Conv_198"> + <dim>64</dim> + <dim>192</dim> + <dim>3</dim> + <dim>3</dim> + </port> + </output> + </layer> + <layer id="48" name="/conv2x_4/conv2x_4.0/Conv/WithoutBiases" type="Convolution" version="opset1"> + <data strides="1, 1" dilations="1, 1" pads_begin="1, 1" pads_end="1, 1" auto_pad="explicit" /> + <input> + <port id="0" precision="FP32"> + <dim>32</dim> + <dim>192</dim> + <dim>32</dim> + <dim>32</dim> + </port> + <port id="1" precision="FP32"> + <dim>64</dim> + <dim>192</dim> + <dim>3</dim> + <dim>3</dim> + </port> + </input> + <output> + <port id="2" precision="FP32"> + <dim>32</dim> + <dim>64</dim> + <dim>32</dim> + <dim>32</dim> + </port> + </output> + </layer> + <layer id="49" name="Reshape_187" type="Const" version="opset1"> + <data element_type="f32" shape="1, 64, 1, 1" offset="1616464" size="256" /> + <output> + <port id="0" precision="FP32"> + <dim>1</dim> + <dim>64</dim> + <dim>1</dim> + <dim>1</dim> + </port> + </output> + </layer> + <layer id="50" name="/conv2x_4/conv2x_4.0/Conv" type="Add" version="opset1"> + <data auto_broadcast="numpy" /> + <input> + <port id="0" precision="FP32"> + <dim>32</dim> + <dim>64</dim> + <dim>32</dim> + <dim>32</dim> + </port> + <port id="1" precision="FP32"> + <dim>1</dim> + <dim>64</dim> + <dim>1</dim> + <dim>1</dim> + </port> + </input> + <output> + <port id="2" precision="FP32" names="/conv2x_4/conv2x_4.0/Conv_output_0"> + <dim>32</dim> + <dim>64</dim> + <dim>32</dim> + <dim>32</dim> + </port> + </output> + </layer> + <layer id="51" name="/conv2x_4/conv2x_4.2/Relu" type="ReLU" version="opset1"> + <input> + <port id="0" precision="FP32"> + <dim>32</dim> + <dim>64</dim> + <dim>32</dim> + <dim>32</dim> + </port> + </input> + <output> + <port id="1" precision="FP32" names="/conv2x_4/conv2x_4.2/Relu_output_0"> + <dim>32</dim> + <dim>64</dim> + <dim>32</dim> + <dim>32</dim> + </port> + </output> + </layer> + <layer id="52" name="onnx::Conv_201" type="Const" version="opset1"> + <data element_type="f32" shape="64, 64, 3, 3" offset="1616720" size="147456" /> + <output> + <port id="0" precision="FP32" names="onnx::Conv_201"> + <dim>64</dim> + <dim>64</dim> + <dim>3</dim> + <dim>3</dim> + </port> + </output> + </layer> + <layer id="53" name="/conv2x_4/conv2x_4.3/Conv/WithoutBiases" type="Convolution" version="opset1"> + <data strides="1, 1" dilations="1, 1" pads_begin="1, 1" pads_end="1, 1" auto_pad="explicit" /> + <input> + <port id="0" precision="FP32"> + <dim>32</dim> + <dim>64</dim> + <dim>32</dim> + <dim>32</dim> + </port> + <port id="1" precision="FP32"> + <dim>64</dim> + <dim>64</dim> + <dim>3</dim> + <dim>3</dim> + </port> + </input> + <output> + <port id="2" precision="FP32"> + <dim>32</dim> + <dim>64</dim> + <dim>32</dim> + <dim>32</dim> + </port> + </output> + </layer> + <layer id="54" name="Reshape_204" type="Const" version="opset1"> + <data element_type="f32" shape="1, 64, 1, 1" offset="1764176" size="256" /> + <output> + <port id="0" precision="FP32"> + <dim>1</dim> + <dim>64</dim> + <dim>1</dim> + <dim>1</dim> + </port> + </output> + </layer> + <layer id="55" name="/conv2x_4/conv2x_4.3/Conv" type="Add" version="opset1"> + <data auto_broadcast="numpy" /> + <input> + <port id="0" precision="FP32"> + <dim>32</dim> + <dim>64</dim> + <dim>32</dim> + <dim>32</dim> + </port> + <port id="1" precision="FP32"> + <dim>1</dim> + <dim>64</dim> + <dim>1</dim> + <dim>1</dim> + </port> + </input> + <output> + <port id="2" precision="FP32" names="/conv2x_4/conv2x_4.3/Conv_output_0"> + <dim>32</dim> + <dim>64</dim> + <dim>32</dim> + <dim>32</dim> + </port> + </output> + </layer> + <layer id="56" name="/conv2x_4/conv2x_4.5/Relu" type="ReLU" version="opset1"> + <input> + <port id="0" precision="FP32"> + <dim>32</dim> + <dim>64</dim> + <dim>32</dim> + <dim>32</dim> + </port> + </input> + <output> + <port id="1" precision="FP32" names="/conv2x_4/conv2x_4.5/Relu_output_0"> + <dim>32</dim> + <dim>64</dim> + <dim>32</dim> + <dim>32</dim> + </port> + </output> + </layer> + <layer id="57" name="/upsample_1/Constant" type="Const" version="opset1"> + <data element_type="f32" shape="4" offset="1174080" size="16" /> + <output> + <port id="0" precision="FP32" names="/upsample_1/Constant_output_0"> + <dim>4</dim> + </port> + </output> + </layer> + <layer id="58" name="/upsample_1/Resize" type="Interpolate" version="opset11"> + <data mode="nearest" shape_calculation_mode="scales" coordinate_transformation_mode="asymmetric" nearest_mode="floor" antialias="false" pads_begin="0, 0, 0, 0" pads_end="0, 0, 0, 0" cube_coeff="-0.75" /> + <input> + <port id="0" precision="FP32"> + <dim>32</dim> + <dim>64</dim> + <dim>32</dim> + <dim>32</dim> + </port> + <port id="1" precision="FP32"> + <dim>4</dim> + </port> + </input> + <output> + <port id="2" precision="FP32" names="/upsample_1/Resize_output_0"> + <dim>32</dim> + <dim>64</dim> + <dim>64</dim> + <dim>64</dim> + </port> + </output> + </layer> + <layer id="59" name="/Concat_1" type="Concat" version="opset1"> + <data axis="1" /> + <input> + <port id="0" precision="FP32"> + <dim>32</dim> + <dim>64</dim> + <dim>64</dim> + <dim>64</dim> + </port> + <port id="1" precision="FP32"> + <dim>32</dim> + <dim>32</dim> + <dim>64</dim> + <dim>64</dim> + </port> + </input> + <output> + <port id="2" precision="FP32" names="/Concat_1_output_0"> + <dim>32</dim> + <dim>96</dim> + <dim>64</dim> + <dim>64</dim> + </port> + </output> + </layer> + <layer id="60" name="onnx::Conv_204" type="Const" version="opset1"> + <data element_type="f32" shape="32, 96, 3, 3" offset="1764432" size="110592" /> + <output> + <port id="0" precision="FP32" names="onnx::Conv_204"> + <dim>32</dim> + <dim>96</dim> + <dim>3</dim> + <dim>3</dim> + </port> + </output> + </layer> + <layer id="61" name="/conv2x_5/conv2x_5.0/Conv/WithoutBiases" type="Convolution" version="opset1"> + <data strides="1, 1" dilations="1, 1" pads_begin="1, 1" pads_end="1, 1" auto_pad="explicit" /> + <input> + <port id="0" precision="FP32"> + <dim>32</dim> + <dim>96</dim> + <dim>64</dim> + <dim>64</dim> + </port> + <port id="1" precision="FP32"> + <dim>32</dim> + <dim>96</dim> + <dim>3</dim> + <dim>3</dim> + </port> + </input> + <output> + <port id="2" precision="FP32"> + <dim>32</dim> + <dim>32</dim> + <dim>64</dim> + <dim>64</dim> + </port> + </output> + </layer> + <layer id="62" name="Reshape_225" type="Const" version="opset1"> + <data element_type="f32" shape="1, 32, 1, 1" offset="1875024" size="128" /> + <output> + <port id="0" precision="FP32"> + <dim>1</dim> + <dim>32</dim> + <dim>1</dim> + <dim>1</dim> + </port> + </output> + </layer> + <layer id="63" name="/conv2x_5/conv2x_5.0/Conv" type="Add" version="opset1"> + <data auto_broadcast="numpy" /> + <input> + <port id="0" precision="FP32"> + <dim>32</dim> + <dim>32</dim> + <dim>64</dim> + <dim>64</dim> + </port> + <port id="1" precision="FP32"> + <dim>1</dim> + <dim>32</dim> + <dim>1</dim> + <dim>1</dim> + </port> + </input> + <output> + <port id="2" precision="FP32" names="/conv2x_5/conv2x_5.0/Conv_output_0"> + <dim>32</dim> + <dim>32</dim> + <dim>64</dim> + <dim>64</dim> + </port> + </output> + </layer> + <layer id="64" name="/conv2x_5/conv2x_5.2/Relu" type="ReLU" version="opset1"> + <input> + <port id="0" precision="FP32"> + <dim>32</dim> + <dim>32</dim> + <dim>64</dim> + <dim>64</dim> + </port> + </input> + <output> + <port id="1" precision="FP32" names="/conv2x_5/conv2x_5.2/Relu_output_0"> + <dim>32</dim> + <dim>32</dim> + <dim>64</dim> + <dim>64</dim> + </port> + </output> + </layer> + <layer id="65" name="onnx::Conv_207" type="Const" version="opset1"> + <data element_type="f32" shape="32, 32, 3, 3" offset="1875152" size="36864" /> + <output> + <port id="0" precision="FP32" names="onnx::Conv_207"> + <dim>32</dim> + <dim>32</dim> + <dim>3</dim> + <dim>3</dim> + </port> + </output> + </layer> + <layer id="66" name="/conv2x_5/conv2x_5.3/Conv/WithoutBiases" type="Convolution" version="opset1"> + <data strides="1, 1" dilations="1, 1" pads_begin="1, 1" pads_end="1, 1" auto_pad="explicit" /> + <input> + <port id="0" precision="FP32"> + <dim>32</dim> + <dim>32</dim> + <dim>64</dim> + <dim>64</dim> + </port> + <port id="1" precision="FP32"> + <dim>32</dim> + <dim>32</dim> + <dim>3</dim> + <dim>3</dim> + </port> + </input> + <output> + <port id="2" precision="FP32"> + <dim>32</dim> + <dim>32</dim> + <dim>64</dim> + <dim>64</dim> + </port> + </output> + </layer> + <layer id="67" name="Reshape_242" type="Const" version="opset1"> + <data element_type="f32" shape="1, 32, 1, 1" offset="1912016" size="128" /> + <output> + <port id="0" precision="FP32"> + <dim>1</dim> + <dim>32</dim> + <dim>1</dim> + <dim>1</dim> + </port> + </output> + </layer> + <layer id="68" name="/conv2x_5/conv2x_5.3/Conv" type="Add" version="opset1"> + <data auto_broadcast="numpy" /> + <input> + <port id="0" precision="FP32"> + <dim>32</dim> + <dim>32</dim> + <dim>64</dim> + <dim>64</dim> + </port> + <port id="1" precision="FP32"> + <dim>1</dim> + <dim>32</dim> + <dim>1</dim> + <dim>1</dim> + </port> + </input> + <output> + <port id="2" precision="FP32" names="/conv2x_5/conv2x_5.3/Conv_output_0"> + <dim>32</dim> + <dim>32</dim> + <dim>64</dim> + <dim>64</dim> + </port> + </output> + </layer> + <layer id="69" name="/conv2x_5/conv2x_5.5/Relu" type="ReLU" version="opset1"> + <input> + <port id="0" precision="FP32"> + <dim>32</dim> + <dim>32</dim> + <dim>64</dim> + <dim>64</dim> + </port> + </input> + <output> + <port id="1" precision="FP32" names="/conv2x_5/conv2x_5.5/Relu_output_0"> + <dim>32</dim> + <dim>32</dim> + <dim>64</dim> + <dim>64</dim> + </port> + </output> + </layer> + <layer id="70" name="/upsample_2/Constant" type="Const" version="opset1"> + <data element_type="f32" shape="4" offset="1174080" size="16" /> + <output> + <port id="0" precision="FP32" names="/upsample_2/Constant_output_0"> + <dim>4</dim> + </port> + </output> + </layer> + <layer id="71" name="/upsample_2/Resize" type="Interpolate" version="opset11"> + <data mode="nearest" shape_calculation_mode="scales" coordinate_transformation_mode="asymmetric" nearest_mode="floor" antialias="false" pads_begin="0, 0, 0, 0" pads_end="0, 0, 0, 0" cube_coeff="-0.75" /> + <input> + <port id="0" precision="FP32"> + <dim>32</dim> + <dim>32</dim> + <dim>64</dim> + <dim>64</dim> + </port> + <port id="1" precision="FP32"> + <dim>4</dim> + </port> + </input> + <output> + <port id="2" precision="FP32" names="/upsample_2/Resize_output_0"> + <dim>32</dim> + <dim>32</dim> + <dim>128</dim> + <dim>128</dim> + </port> + </output> + </layer> + <layer id="72" name="/Concat_2" type="Concat" version="opset1"> + <data axis="1" /> + <input> + <port id="0" precision="FP32"> + <dim>32</dim> + <dim>32</dim> + <dim>128</dim> + <dim>128</dim> + </port> + <port id="1" precision="FP32"> + <dim>32</dim> + <dim>16</dim> + <dim>128</dim> + <dim>128</dim> + </port> + </input> + <output> + <port id="2" precision="FP32" names="/Concat_2_output_0"> + <dim>32</dim> + <dim>48</dim> + <dim>128</dim> + <dim>128</dim> + </port> + </output> + </layer> + <layer id="73" name="onnx::Conv_210" type="Const" version="opset1"> + <data element_type="f32" shape="16, 48, 3, 3" offset="1912144" size="27648" /> + <output> + <port id="0" precision="FP32" names="onnx::Conv_210"> + <dim>16</dim> + <dim>48</dim> + <dim>3</dim> + <dim>3</dim> + </port> + </output> + </layer> + <layer id="74" name="/conv2x_6/conv2x_6.0/Conv/WithoutBiases" type="Convolution" version="opset1"> + <data strides="1, 1" dilations="1, 1" pads_begin="1, 1" pads_end="1, 1" auto_pad="explicit" /> + <input> + <port id="0" precision="FP32"> + <dim>32</dim> + <dim>48</dim> + <dim>128</dim> + <dim>128</dim> + </port> + <port id="1" precision="FP32"> + <dim>16</dim> + <dim>48</dim> + <dim>3</dim> + <dim>3</dim> + </port> + </input> + <output> + <port id="2" precision="FP32"> + <dim>32</dim> + <dim>16</dim> + <dim>128</dim> + <dim>128</dim> + </port> + </output> + </layer> + <layer id="75" name="Reshape_263" type="Const" version="opset1"> + <data element_type="f32" shape="1, 16, 1, 1" offset="1939792" size="64" /> + <output> + <port id="0" precision="FP32"> + <dim>1</dim> + <dim>16</dim> + <dim>1</dim> + <dim>1</dim> + </port> + </output> + </layer> + <layer id="76" name="/conv2x_6/conv2x_6.0/Conv" type="Add" version="opset1"> + <data auto_broadcast="numpy" /> + <input> + <port id="0" precision="FP32"> + <dim>32</dim> + <dim>16</dim> + <dim>128</dim> + <dim>128</dim> + </port> + <port id="1" precision="FP32"> + <dim>1</dim> + <dim>16</dim> + <dim>1</dim> + <dim>1</dim> + </port> + </input> + <output> + <port id="2" precision="FP32" names="/conv2x_6/conv2x_6.0/Conv_output_0"> + <dim>32</dim> + <dim>16</dim> + <dim>128</dim> + <dim>128</dim> + </port> + </output> + </layer> + <layer id="77" name="/conv2x_6/conv2x_6.2/Relu" type="ReLU" version="opset1"> + <input> + <port id="0" precision="FP32"> + <dim>32</dim> + <dim>16</dim> + <dim>128</dim> + <dim>128</dim> + </port> + </input> + <output> + <port id="1" precision="FP32" names="/conv2x_6/conv2x_6.2/Relu_output_0"> + <dim>32</dim> + <dim>16</dim> + <dim>128</dim> + <dim>128</dim> + </port> + </output> + </layer> + <layer id="78" name="onnx::Conv_213" type="Const" version="opset1"> + <data element_type="f32" shape="16, 16, 3, 3" offset="1939856" size="9216" /> + <output> + <port id="0" precision="FP32" names="onnx::Conv_213"> + <dim>16</dim> + <dim>16</dim> + <dim>3</dim> + <dim>3</dim> + </port> + </output> + </layer> + <layer id="79" name="/conv2x_6/conv2x_6.3/Conv/WithoutBiases" type="Convolution" version="opset1"> + <data strides="1, 1" dilations="1, 1" pads_begin="1, 1" pads_end="1, 1" auto_pad="explicit" /> + <input> + <port id="0" precision="FP32"> + <dim>32</dim> + <dim>16</dim> + <dim>128</dim> + <dim>128</dim> + </port> + <port id="1" precision="FP32"> + <dim>16</dim> + <dim>16</dim> + <dim>3</dim> + <dim>3</dim> + </port> + </input> + <output> + <port id="2" precision="FP32"> + <dim>32</dim> + <dim>16</dim> + <dim>128</dim> + <dim>128</dim> + </port> + </output> + </layer> + <layer id="80" name="Reshape_280" type="Const" version="opset1"> + <data element_type="f32" shape="1, 16, 1, 1" offset="1949072" size="64" /> + <output> + <port id="0" precision="FP32"> + <dim>1</dim> + <dim>16</dim> + <dim>1</dim> + <dim>1</dim> + </port> + </output> + </layer> + <layer id="81" name="/conv2x_6/conv2x_6.3/Conv" type="Add" version="opset1"> + <data auto_broadcast="numpy" /> + <input> + <port id="0" precision="FP32"> + <dim>32</dim> + <dim>16</dim> + <dim>128</dim> + <dim>128</dim> + </port> + <port id="1" precision="FP32"> + <dim>1</dim> + <dim>16</dim> + <dim>1</dim> + <dim>1</dim> + </port> + </input> + <output> + <port id="2" precision="FP32" names="/conv2x_6/conv2x_6.3/Conv_output_0"> + <dim>32</dim> + <dim>16</dim> + <dim>128</dim> + <dim>128</dim> + </port> + </output> + </layer> + <layer id="82" name="/conv2x_6/conv2x_6.5/Relu" type="ReLU" version="opset1"> + <input> + <port id="0" precision="FP32"> + <dim>32</dim> + <dim>16</dim> + <dim>128</dim> + <dim>128</dim> + </port> + </input> + <output> + <port id="1" precision="FP32" names="/conv2x_6/conv2x_6.5/Relu_output_0"> + <dim>32</dim> + <dim>16</dim> + <dim>128</dim> + <dim>128</dim> + </port> + </output> + </layer> + <layer id="83" name="out.weight" type="Const" version="opset1"> + <data element_type="f32" shape="1, 16, 1, 1" offset="1949136" size="64" /> + <output> + <port id="0" precision="FP32" names="out.weight"> + <dim>1</dim> + <dim>16</dim> + <dim>1</dim> + <dim>1</dim> + </port> + </output> + </layer> + <layer id="84" name="/out/Conv/WithoutBiases" type="Convolution" version="opset1"> + <data strides="1, 1" dilations="1, 1" pads_begin="0, 0" pads_end="0, 0" auto_pad="explicit" /> + <input> + <port id="0" precision="FP32"> + <dim>32</dim> + <dim>16</dim> + <dim>128</dim> + <dim>128</dim> + </port> + <port id="1" precision="FP32"> + <dim>1</dim> + <dim>16</dim> + <dim>1</dim> + <dim>1</dim> + </port> + </input> + <output> + <port id="2" precision="FP32"> + <dim>32</dim> + <dim>1</dim> + <dim>128</dim> + <dim>128</dim> + </port> + </output> + </layer> + <layer id="85" name="Reshape_297" type="Const" version="opset1"> + <data element_type="f32" shape="1, 1, 1, 1" offset="1949200" size="4" /> + <output> + <port id="0" precision="FP32"> + <dim>1</dim> + <dim>1</dim> + <dim>1</dim> + <dim>1</dim> + </port> + </output> + </layer> + <layer id="86" name="/out/Conv" type="Add" version="opset1"> + <data auto_broadcast="numpy" /> + <input> + <port id="0" precision="FP32"> + <dim>32</dim> + <dim>1</dim> + <dim>128</dim> + <dim>128</dim> + </port> + <port id="1" precision="FP32"> + <dim>1</dim> + <dim>1</dim> + <dim>1</dim> + <dim>1</dim> + </port> + </input> + <output> + <port id="2" precision="FP32" names="/out/Conv_output_0"> + <dim>32</dim> + <dim>1</dim> + <dim>128</dim> + <dim>128</dim> + </port> + </output> + </layer> + <layer id="87" name="Constant_1585" type="Const" version="opset1"> + <data element_type="i64" shape="2" offset="1949204" size="16" /> + <output> + <port id="0" precision="I64"> + <dim>2</dim> + </port> + </output> + </layer> + <layer id="88" name="Constant_1588" type="Const" version="opset1"> + <data element_type="i64" shape="2" offset="1949220" size="16" /> + <output> + <port id="0" precision="I64"> + <dim>2</dim> + </port> + </output> + </layer> + <layer id="89" name="Constant_1591" type="Const" version="opset1"> + <data element_type="i64" shape="2" offset="1949236" size="16" /> + <output> + <port id="0" precision="I64"> + <dim>2</dim> + </port> + </output> + </layer> + <layer id="90" name="/Slice" type="StridedSlice" version="opset1"> + <data begin_mask="1, 0" end_mask="1, 0" new_axis_mask="" shrink_axis_mask="" ellipsis_mask="" /> + <input> + <port id="0" precision="FP32"> + <dim>32</dim> + <dim>3</dim> + <dim>128</dim> + <dim>128</dim> + </port> + <port id="1" precision="I64"> + <dim>2</dim> + </port> + <port id="2" precision="I64"> + <dim>2</dim> + </port> + <port id="3" precision="I64"> + <dim>2</dim> + </port> + </input> + <output> + <port id="4" precision="FP32" names="/Slice_output_0"> + <dim>32</dim> + <dim>1</dim> + <dim>128</dim> + <dim>128</dim> + </port> + </output> + </layer> + <layer id="91" name="172" type="Add" version="opset1"> + <data auto_broadcast="numpy" /> + <input> + <port id="0" precision="FP32"> + <dim>32</dim> + <dim>1</dim> + <dim>128</dim> + <dim>128</dim> + </port> + <port id="1" precision="FP32"> + <dim>32</dim> + <dim>1</dim> + <dim>128</dim> + <dim>128</dim> + </port> + </input> + <output> + <port id="2" precision="FP32" names="172"> + <dim>32</dim> + <dim>1</dim> + <dim>128</dim> + <dim>128</dim> + </port> + </output> + </layer> + <layer id="92" name="172/sink_port_0" type="Result" version="opset1"> + <input> + <port id="0" precision="FP32"> + <dim>32</dim> + <dim>1</dim> + <dim>128</dim> + <dim>128</dim> + </port> + </input> + </layer> + </layers> + <edges> + <edge from-layer="0" from-port="0" to-layer="2" to-port="0" /> + <edge from-layer="0" from-port="0" to-layer="90" to-port="0" /> + <edge from-layer="1" from-port="0" to-layer="2" to-port="1" /> + <edge from-layer="2" from-port="2" to-layer="4" to-port="0" /> + <edge from-layer="3" from-port="0" to-layer="4" to-port="1" /> + <edge from-layer="4" from-port="2" to-layer="5" to-port="0" /> + <edge from-layer="5" from-port="1" to-layer="7" to-port="0" /> + <edge from-layer="6" from-port="0" to-layer="7" to-port="1" /> + <edge from-layer="7" from-port="2" to-layer="9" to-port="0" /> + <edge from-layer="8" from-port="0" to-layer="9" to-port="1" /> + <edge from-layer="9" from-port="2" to-layer="10" to-port="0" /> + <edge from-layer="10" from-port="1" to-layer="11" to-port="0" /> + <edge from-layer="10" from-port="1" to-layer="72" to-port="1" /> + <edge from-layer="11" from-port="1" to-layer="13" to-port="0" /> + <edge from-layer="12" from-port="0" to-layer="13" to-port="1" /> + <edge from-layer="13" from-port="2" to-layer="15" to-port="0" /> + <edge from-layer="14" from-port="0" to-layer="15" to-port="1" /> + <edge from-layer="15" from-port="2" to-layer="16" to-port="0" /> + <edge from-layer="16" from-port="1" to-layer="18" to-port="0" /> + <edge from-layer="17" from-port="0" to-layer="18" to-port="1" /> + <edge from-layer="18" from-port="2" to-layer="20" to-port="0" /> + <edge from-layer="19" from-port="0" to-layer="20" to-port="1" /> + <edge from-layer="20" from-port="2" to-layer="21" to-port="0" /> + <edge from-layer="21" from-port="1" to-layer="59" to-port="1" /> + <edge from-layer="21" from-port="1" to-layer="22" to-port="0" /> + <edge from-layer="22" from-port="1" to-layer="24" to-port="0" /> + <edge from-layer="23" from-port="0" to-layer="24" to-port="1" /> + <edge from-layer="24" from-port="2" to-layer="26" to-port="0" /> + <edge from-layer="25" from-port="0" to-layer="26" to-port="1" /> + <edge from-layer="26" from-port="2" to-layer="27" to-port="0" /> + <edge from-layer="27" from-port="1" to-layer="29" to-port="0" /> + <edge from-layer="28" from-port="0" to-layer="29" to-port="1" /> + <edge from-layer="29" from-port="2" to-layer="31" to-port="0" /> + <edge from-layer="30" from-port="0" to-layer="31" to-port="1" /> + <edge from-layer="31" from-port="2" to-layer="32" to-port="0" /> + <edge from-layer="32" from-port="1" to-layer="46" to-port="1" /> + <edge from-layer="32" from-port="1" to-layer="33" to-port="0" /> + <edge from-layer="33" from-port="1" to-layer="35" to-port="0" /> + <edge from-layer="34" from-port="0" to-layer="35" to-port="1" /> + <edge from-layer="35" from-port="2" to-layer="37" to-port="0" /> + <edge from-layer="36" from-port="0" to-layer="37" to-port="1" /> + <edge from-layer="37" from-port="2" to-layer="38" to-port="0" /> + <edge from-layer="38" from-port="1" to-layer="40" to-port="0" /> + <edge from-layer="39" from-port="0" to-layer="40" to-port="1" /> + <edge from-layer="40" from-port="2" to-layer="42" to-port="0" /> + <edge from-layer="41" from-port="0" to-layer="42" to-port="1" /> + <edge from-layer="42" from-port="2" to-layer="43" to-port="0" /> + <edge from-layer="43" from-port="1" to-layer="45" to-port="0" /> + <edge from-layer="44" from-port="0" to-layer="45" to-port="1" /> + <edge from-layer="45" from-port="2" to-layer="46" to-port="0" /> + <edge from-layer="46" from-port="2" to-layer="48" to-port="0" /> + <edge from-layer="47" from-port="0" to-layer="48" to-port="1" /> + <edge from-layer="48" from-port="2" to-layer="50" to-port="0" /> + <edge from-layer="49" from-port="0" to-layer="50" to-port="1" /> + <edge from-layer="50" from-port="2" to-layer="51" to-port="0" /> + <edge from-layer="51" from-port="1" to-layer="53" to-port="0" /> + <edge from-layer="52" from-port="0" to-layer="53" to-port="1" /> + <edge from-layer="53" from-port="2" to-layer="55" to-port="0" /> + <edge from-layer="54" from-port="0" to-layer="55" to-port="1" /> + <edge from-layer="55" from-port="2" to-layer="56" to-port="0" /> + <edge from-layer="56" from-port="1" to-layer="58" to-port="0" /> + <edge from-layer="57" from-port="0" to-layer="58" to-port="1" /> + <edge from-layer="58" from-port="2" to-layer="59" to-port="0" /> + <edge from-layer="59" from-port="2" to-layer="61" to-port="0" /> + <edge from-layer="60" from-port="0" to-layer="61" to-port="1" /> + <edge from-layer="61" from-port="2" to-layer="63" to-port="0" /> + <edge from-layer="62" from-port="0" to-layer="63" to-port="1" /> + <edge from-layer="63" from-port="2" to-layer="64" to-port="0" /> + <edge from-layer="64" from-port="1" to-layer="66" to-port="0" /> + <edge from-layer="65" from-port="0" to-layer="66" to-port="1" /> + <edge from-layer="66" from-port="2" to-layer="68" to-port="0" /> + <edge from-layer="67" from-port="0" to-layer="68" to-port="1" /> + <edge from-layer="68" from-port="2" to-layer="69" to-port="0" /> + <edge from-layer="69" from-port="1" to-layer="71" to-port="0" /> + <edge from-layer="70" from-port="0" to-layer="71" to-port="1" /> + <edge from-layer="71" from-port="2" to-layer="72" to-port="0" /> + <edge from-layer="72" from-port="2" to-layer="74" to-port="0" /> + <edge from-layer="73" from-port="0" to-layer="74" to-port="1" /> + <edge from-layer="74" from-port="2" to-layer="76" to-port="0" /> + <edge from-layer="75" from-port="0" to-layer="76" to-port="1" /> + <edge from-layer="76" from-port="2" to-layer="77" to-port="0" /> + <edge from-layer="77" from-port="1" to-layer="79" to-port="0" /> + <edge from-layer="78" from-port="0" to-layer="79" to-port="1" /> + <edge from-layer="79" from-port="2" to-layer="81" to-port="0" /> + <edge from-layer="80" from-port="0" to-layer="81" to-port="1" /> + <edge from-layer="81" from-port="2" to-layer="82" to-port="0" /> + <edge from-layer="82" from-port="1" to-layer="84" to-port="0" /> + <edge from-layer="83" from-port="0" to-layer="84" to-port="1" /> + <edge from-layer="84" from-port="2" to-layer="86" to-port="0" /> + <edge from-layer="85" from-port="0" to-layer="86" to-port="1" /> + <edge from-layer="86" from-port="2" to-layer="91" to-port="0" /> + <edge from-layer="87" from-port="0" to-layer="90" to-port="1" /> + <edge from-layer="88" from-port="0" to-layer="90" to-port="2" /> + <edge from-layer="89" from-port="0" to-layer="90" to-port="3" /> + <edge from-layer="90" from-port="4" to-layer="91" to-port="1" /> + <edge from-layer="91" from-port="2" to-layer="92" to-port="0" /> + </edges> + <rt_info> + <MO_version value="2024.6.0-17404-4c0f47d2335-releases/2024/6" /> + <Runtime_version value="2023.3.0-13775-ceeafaf64f3-releases/2023/3" /> + <conversion_parameters> + <input_model value="DIR/model_bmp.onnx" /> + <is_python_api_used value="False" /> + </conversion_parameters> + <legacy_frontend value="False" /> + </rt_info> +</net> diff --git a/python/openvino/demo/models/model_bmp_de5a_area.bin b/python/openvino/demo/models/model_bmp_de5a_area.bin Binary files differnew file mode 100644 index 0000000..f1b6f55 --- /dev/null +++ b/python/openvino/demo/models/model_bmp_de5a_area.bin |
