sirbastiano94 commited on
Commit
1a7ae11
·
verified ·
1 Parent(s): aa10b5b

Add files using upload-large-folder tool

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. comprehensive_distillation_20251112_013111/experiments_summary.json +1 -0
  2. exp_1762448560_9364_s96_f16_d3_m2_3_3_4/convert_openvino.sh +30 -0
  3. exp_1762448560_9364_s96_f16_d3_m2_3_3_4/experiment.log +100 -0
  4. exp_1762448560_9364_s96_f16_d3_m2_3_3_4/model_info.json +140 -0
  5. exp_1762448560_9364_s96_f16_d3_m2_3_3_4/pipeline_results.json +140 -0
  6. exp_1762448560_9364_s96_f16_d3_m2_3_3_4/run_inference.sh +19 -0
  7. exp_1762448567_7023_s96_f18_d3_m1_1_1_1/convert_openvino.sh +30 -0
  8. exp_1762448567_7023_s96_f18_d3_m1_1_1_1/experiment.log +98 -0
  9. exp_1762448567_7023_s96_f18_d3_m1_1_1_1/model_info.json +138 -0
  10. exp_1762448567_7023_s96_f18_d3_m1_1_1_1/pipeline_results.json +138 -0
  11. exp_1762448567_7023_s96_f18_d3_m1_1_1_1/run_inference.sh +19 -0
  12. exp_1762448652_9207_s96_f20_d3_m2_2_2_2/convert_openvino.sh +30 -0
  13. exp_1762448652_9207_s96_f20_d3_m2_2_2_2/experiment.log +98 -0
  14. exp_1762448652_9207_s96_f20_d3_m2_2_2_2/model_info.json +138 -0
  15. exp_1762448652_9207_s96_f20_d3_m2_2_2_2/pipeline_results.json +138 -0
  16. exp_1762448652_9207_s96_f20_d3_m2_2_2_2/run_inference.sh +19 -0
  17. exp_1762448723_5984_s96_f22_d3_m1_1_1_1/convert_openvino.sh +30 -0
  18. exp_1762448723_5984_s96_f22_d3_m1_1_1_1/experiment.log +98 -0
  19. exp_1762448723_5984_s96_f22_d3_m1_1_1_1/model_info.json +138 -0
  20. exp_1762448723_5984_s96_f22_d3_m1_1_1_1/pipeline_results.json +138 -0
  21. exp_1762448723_5984_s96_f22_d3_m1_1_1_1/run_inference.sh +19 -0
  22. exp_1762448809_7498_s96_f24_d3_m2_2_2_2/convert_openvino.sh +30 -0
  23. exp_1762448809_7498_s96_f24_d3_m2_2_2_2/experiment.log +98 -0
  24. exp_1762448809_7498_s96_f24_d3_m2_2_2_2/model_info.json +138 -0
  25. exp_1762448809_7498_s96_f24_d3_m2_2_2_2/pipeline_results.json +138 -0
  26. exp_1762448809_7498_s96_f24_d3_m2_2_2_2/run_inference.sh +19 -0
  27. exp_1762448858_8082_s96_f24_d3_m1_2_3_4/convert_openvino.sh +30 -0
  28. exp_1762448858_8082_s96_f24_d3_m1_2_3_4/experiment.log +101 -0
  29. exp_1762448858_8082_s96_f24_d3_m1_2_3_4/model_info.json +141 -0
  30. exp_1762448858_8082_s96_f24_d3_m1_2_3_4/pipeline_results.json +141 -0
  31. exp_1762448858_8082_s96_f24_d3_m1_2_3_4/run_inference.sh +19 -0
  32. exp_1762448908_9161_s96_f26_d3_m2_2_2_3/convert_openvino.sh +30 -0
  33. exp_1762448908_9161_s96_f26_d3_m2_2_2_3/experiment.log +99 -0
  34. exp_1762448908_9161_s96_f26_d3_m2_2_2_3/model_info.json +139 -0
  35. exp_1762448908_9161_s96_f26_d3_m2_2_2_3/pipeline_results.json +139 -0
  36. exp_1762448908_9161_s96_f26_d3_m2_2_2_3/run_inference.sh +19 -0
  37. exp_1762449009_9308_s96_f28_d3_m1_2_2_3/convert_openvino.sh +30 -0
  38. exp_1762449009_9308_s96_f28_d3_m1_2_2_3/experiment.log +100 -0
  39. exp_1762449009_9308_s96_f28_d3_m1_2_2_3/model_info.json +140 -0
  40. exp_1762449009_9308_s96_f28_d3_m1_2_2_3/pipeline_results.json +140 -0
  41. exp_1762449009_9308_s96_f28_d3_m1_2_2_3/run_inference.sh +19 -0
  42. exp_1762449060_9675_s96_f30_d3_m1_2_2_3/convert_openvino.sh +30 -0
  43. exp_1762449060_9675_s96_f30_d3_m1_2_2_3/experiment.log +100 -0
  44. exp_1762449060_9675_s96_f30_d3_m1_2_2_3/model_info.json +140 -0
  45. exp_1762449060_9675_s96_f30_d3_m1_2_2_3/pipeline_results.json +140 -0
  46. exp_1762449060_9675_s96_f30_d3_m1_2_2_3/run_inference.sh +19 -0
  47. exp_1762449133_5525_s96_f32_d3_m1_2_2_2/convert_openvino.sh +30 -0
  48. exp_1762449133_5525_s96_f32_d3_m1_2_2_2/experiment.log +99 -0
  49. exp_1762449133_5525_s96_f32_d3_m1_2_2_2/model_info.json +139 -0
  50. exp_1762449133_5525_s96_f32_d3_m1_2_2_2/pipeline_results.json +139 -0
comprehensive_distillation_20251112_013111/experiments_summary.json ADDED
@@ -0,0 +1 @@
 
 
1
+ []
exp_1762448560_9364_s96_f16_d3_m2_3_3_4/convert_openvino.sh ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #!/bin/bash
2
+ # Source OpenVINO environment
3
+ source /opt/intel/openvino/bin/setupvars.sh
4
+
5
+ # OpenVINO Model Optimizer script for ONNX model
6
+ PYPATH=/opt/intel/openvino_2020.3.194/deployment_tools/model_optimizer/mo.py
7
+ ONNX_MODEL=/home/mount/experiments/exp_1762448560_9364_s96_f16_d3_m2_3_3_4/onnx/model.onnx
8
+ OUTPUT_DIR=/home/mount/experiments/exp_1762448560_9364_s96_f16_d3_m2_3_3_4/openvino
9
+
10
+ echo "Converting ONNX model to OpenVINO IR format..."
11
+ echo "Input model: $ONNX_MODEL"
12
+ echo "Output directory: $OUTPUT_DIR"
13
+
14
+ # Create output directory if it doesn't exist
15
+ mkdir -p "$OUTPUT_DIR"
16
+
17
+ python3 $PYPATH \
18
+ --input_model "$ONNX_MODEL" \
19
+ --data_type FP16 \
20
+ --input_shape "[1,8,96,96]" \
21
+ --mean_values "[0,0,0,0,0,0,0,0]" \
22
+ --scale_values "[1,1,1,1,1,1,1,1]" \
23
+ --progress \
24
+ --stream_output \
25
+ --output_dir "$OUTPUT_DIR" \
26
+ --model_name model
27
+
28
+ echo "OpenVINO conversion completed!"
29
+ echo "Generated files in $OUTPUT_DIR:"
30
+ ls -la "$OUTPUT_DIR/"
exp_1762448560_9364_s96_f16_d3_m2_3_3_4/experiment.log ADDED
@@ -0,0 +1,100 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2025-11-06 18:02:40 - INFO - === Model Creation Phase ===
2
+ 2025-11-06 18:02:40 - INFO - Creating UNet model with config: {'input_size': 96, 'base_filters': 16, 'depth': 3, 'channel_multipliers': [2, 3, 3, 4], 'n_channels': 8, 'n_classes': 3}
3
+ 2025-11-06 18:02:40 - INFO - Model created successfully:
4
+ 2025-11-06 18:02:40 - INFO - Depth: 3
5
+ 2025-11-06 18:02:40 - INFO - Channels per level: [32, 48, 48, 64]
6
+ 2025-11-06 18:02:40 - INFO - Total parameters: 194,195
7
+ 2025-11-06 18:02:40 - INFO - Parameter memory: 0.74 MB
8
+ 2025-11-06 18:02:40 - INFO - === Detailed Architecture Analysis ===
9
+ 2025-11-06 18:02:40 - INFO - Input vector dimension: 73,728 (8 × 96 × 96)
10
+ 2025-11-06 18:02:40 - INFO - Component-wise parameter breakdown:
11
+ 2025-11-06 18:02:40 - INFO - Encoder - encoders.0.channel_proj: 288 parameters
12
+ 2025-11-06 18:02:40 - INFO - Encoder - encoders.0.convnext_block.dwconv: 1,600 parameters
13
+ 2025-11-06 18:02:40 - INFO - Encoder - encoders.0.convnext_block.norm: 64 parameters
14
+ 2025-11-06 18:02:40 - INFO - Encoder - encoders.0.convnext_block.pwconv1: 4,224 parameters
15
+ 2025-11-06 18:02:40 - INFO - Encoder - encoders.0.convnext_block.pwconv2: 4,128 parameters
16
+ 2025-11-06 18:02:40 - INFO - Encoder - encoders.1.channel_proj: 1,584 parameters
17
+ 2025-11-06 18:02:40 - INFO - Encoder - encoders.1.convnext_block.dwconv: 2,400 parameters
18
+ 2025-11-06 18:02:40 - INFO - Encoder - encoders.1.convnext_block.norm: 96 parameters
19
+ 2025-11-06 18:02:40 - INFO - Encoder - encoders.1.convnext_block.pwconv1: 9,408 parameters
20
+ 2025-11-06 18:02:40 - INFO - Encoder - encoders.1.convnext_block.pwconv2: 9,264 parameters
21
+ 2025-11-06 18:02:40 - INFO - Encoder - encoders.2.convnext_block.dwconv: 2,400 parameters
22
+ 2025-11-06 18:02:40 - INFO - Encoder - encoders.2.convnext_block.norm: 96 parameters
23
+ 2025-11-06 18:02:40 - INFO - Encoder - encoders.2.convnext_block.pwconv1: 9,408 parameters
24
+ 2025-11-06 18:02:40 - INFO - Encoder - encoders.2.convnext_block.pwconv2: 9,264 parameters
25
+ 2025-11-06 18:02:40 - INFO - Bottleneck - bottleneck.channel_proj: 3,136 parameters
26
+ 2025-11-06 18:02:40 - INFO - Bottleneck - bottleneck.convnext_block.dwconv: 3,200 parameters
27
+ 2025-11-06 18:02:40 - INFO - Bottleneck - bottleneck.convnext_block.norm: 128 parameters
28
+ 2025-11-06 18:02:40 - INFO - Bottleneck - bottleneck.convnext_block.pwconv1: 16,640 parameters
29
+ 2025-11-06 18:02:40 - INFO - Bottleneck - bottleneck.convnext_block.pwconv2: 16,448 parameters
30
+ 2025-11-06 18:02:40 - INFO - Decoder - upsamplers.0: 16,448 parameters
31
+ 2025-11-06 18:02:40 - INFO - Decoder - upsamplers.1: 9,264 parameters
32
+ 2025-11-06 18:02:40 - INFO - Decoder - upsamplers.2: 9,264 parameters
33
+ 2025-11-06 18:02:40 - INFO - Decoder - decoders.0.channel_proj: 5,424 parameters
34
+ 2025-11-06 18:02:40 - INFO - Decoder - decoders.0.convnext_block.dwconv: 2,400 parameters
35
+ 2025-11-06 18:02:40 - INFO - Decoder - decoders.0.convnext_block.norm: 96 parameters
36
+ 2025-11-06 18:02:40 - INFO - Decoder - decoders.0.convnext_block.pwconv1: 9,408 parameters
37
+ 2025-11-06 18:02:40 - INFO - Decoder - decoders.0.convnext_block.pwconv2: 9,264 parameters
38
+ 2025-11-06 18:02:40 - INFO - Decoder - decoders.1.channel_proj: 4,656 parameters
39
+ 2025-11-06 18:02:40 - INFO - Decoder - decoders.1.convnext_block.dwconv: 2,400 parameters
40
+ 2025-11-06 18:02:40 - INFO - Decoder - decoders.1.convnext_block.norm: 96 parameters
41
+ 2025-11-06 18:02:40 - INFO - Decoder - decoders.1.convnext_block.pwconv1: 9,408 parameters
42
+ 2025-11-06 18:02:40 - INFO - Decoder - decoders.1.convnext_block.pwconv2: 9,264 parameters
43
+ 2025-11-06 18:02:40 - INFO - Decoder - decoders.2.channel_proj: 2,592 parameters
44
+ 2025-11-06 18:02:40 - INFO - Decoder - decoders.2.convnext_block.dwconv: 1,600 parameters
45
+ 2025-11-06 18:02:40 - INFO - Decoder - decoders.2.convnext_block.norm: 64 parameters
46
+ 2025-11-06 18:02:40 - INFO - Decoder - decoders.2.convnext_block.pwconv1: 4,224 parameters
47
+ 2025-11-06 18:02:40 - INFO - Decoder - decoders.2.convnext_block.pwconv2: 4,128 parameters
48
+ 2025-11-06 18:02:40 - INFO - Other - final_conv: 99 parameters
49
+ 2025-11-06 18:02:40 - INFO - Parameter distribution summary:
50
+ 2025-11-06 18:02:40 - INFO - Encoder parameters: 54,224 (28.0%)
51
+ 2025-11-06 18:02:40 - INFO - Decoder parameters: 100,000 (51.6%)
52
+ 2025-11-06 18:02:40 - INFO - Bottleneck parameters: 39,552 (20.4%)
53
+ 2025-11-06 18:02:40 - INFO - Other parameters: 99 (0.1%)
54
+ 2025-11-06 18:02:40 - INFO - Latent space dimensions (feature maps at each level):
55
+ 2025-11-06 18:02:40 - INFO - Level 0: 32 × 96 × 96 = 294,912 elements
56
+ 2025-11-06 18:02:40 - INFO - Level 1: 48 × 48 × 48 = 110,592 elements
57
+ 2025-11-06 18:02:40 - INFO - Level 2: 48 × 24 × 24 = 27,648 elements
58
+ 2025-11-06 18:02:40 - INFO - Level 3: 64 × 12 × 12 = 9,216 elements
59
+ 2025-11-06 18:02:40 - INFO - Skip connection dimensions:
60
+ 2025-11-06 18:02:40 - INFO - Skip 0: 32 × 96 × 96 = 294,912 elements
61
+ 2025-11-06 18:02:40 - INFO - Skip 1: 48 × 48 × 48 = 110,592 elements
62
+ 2025-11-06 18:02:40 - INFO - Skip 2: 48 × 24 × 24 = 27,648 elements
63
+ 2025-11-06 18:02:40 - INFO - Memory analysis:
64
+ 2025-11-06 18:02:40 - INFO - Peak feature map memory (inference): 1.97 MB
65
+ 2025-11-06 18:02:40 - INFO - Peak feature map memory (training): 3.94 MB (with gradients)
66
+ 2025-11-06 18:02:40 - INFO - Output vector dimension: 27,648 (3 × 96 × 96)
67
+ 2025-11-06 18:02:40 - INFO - PyTorch model saved to: /home/philab/Desktop/hydranet/experiments/exp_1762448560_9364_s96_f16_d3_m2_3_3_4/pytorch/model.pt
68
+ 2025-11-06 18:02:40 - INFO - === ONNX Conversion Phase ===
69
+ 2025-11-06 18:02:40 - INFO - === Model Export Diagnostics ===
70
+ 2025-11-06 18:02:40 - INFO - PyTorch version: 1.9.0+cu102
71
+ 2025-11-06 18:02:40 - INFO - Model parameters: 194,195
72
+ 2025-11-06 18:02:40 - INFO - Model memory: 0.74 MB
73
+ 2025-11-06 18:02:40 - INFO - Starting ONNX export with opset version 11
74
+ 2025-11-06 18:02:40 - INFO - Model input shape: torch.Size([1, 8, 96, 96])
75
+ 2025-11-06 18:02:40 - INFO - Model input dtype: torch.float32
76
+ 2025-11-06 18:02:40 - INFO - Forward pass successful. Output shape: torch.Size([1, 3, 96, 96])
77
+ 2025-11-06 18:02:40 - INFO - Output dtype: torch.float32
78
+ 2025-11-06 18:02:40 - INFO - Output value range: [-0.3959, 0.6202]
79
+ 2025-11-06 18:02:40 - INFO - Model successfully exported to /home/philab/Desktop/hydranet/experiments/exp_1762448560_9364_s96_f16_d3_m2_3_3_4/onnx/model.onnx
80
+ 2025-11-06 18:02:40 - INFO - ONNX model size: 0.75 MB
81
+ 2025-11-06 18:02:40 - INFO - Saved dummy input with shape (1, 8, 96, 96) to /home/philab/Desktop/hydranet/experiments/exp_1762448560_9364_s96_f16_d3_m2_3_3_4/onnx/sample_input.npy
82
+ 2025-11-06 18:02:40 - INFO - Input data type: float32
83
+ 2025-11-06 18:02:40 - INFO - Input value range: [-4.4606, 4.2294]
84
+ 2025-11-06 18:02:40 - INFO - === OpenVINO Conversion Phase ===
85
+ 2025-11-06 18:02:40 - INFO - Starting OpenVINO conversion in Docker container...
86
+ 2025-11-06 18:02:44 - INFO - OpenVINO conversion completed in 4.09 seconds
87
+ 2025-11-06 18:02:44 - INFO - OpenVINO model files created:
88
+ 2025-11-06 18:02:44 - INFO - XML file: /home/philab/Desktop/hydranet/experiments/exp_1762448560_9364_s96_f16_d3_m2_3_3_4/openvino/model.xml (0.09 MB)
89
+ 2025-11-06 18:02:44 - INFO - BIN file: /home/philab/Desktop/hydranet/experiments/exp_1762448560_9364_s96_f16_d3_m2_3_3_4/openvino/model.bin (0.37 MB)
90
+ 2025-11-06 18:02:44 - INFO - === Myriad Inference Phase ===
91
+ 2025-11-06 18:02:44 - INFO - Starting Myriad inference in Docker container...
92
+ 2025-11-06 18:02:47 - INFO - Myriad inference completed in 2.74 seconds
93
+ 2025-11-06 18:02:47 - INFO - Actual inference time: 0.132681 seconds
94
+ 2025-11-06 18:02:47 - INFO - ✅ Complete pipeline executed successfully!
95
+ 2025-11-06 18:02:47 - INFO - ✅ Experiment 11 completed successfully
96
+ 2025-11-06 18:02:47 - INFO - Inference time: 0.132681s
97
+ 2025-11-06 18:02:47 - INFO -
98
+ === Experiment 12/2475 ===
99
+ 2025-11-06 18:02:47 - INFO - Experiment ID: exp_1762448567_7023_s96_f18_d3_m1_1_1_1
100
+ 2025-11-06 18:02:47 - INFO - Experiment directory: /home/philab/Desktop/hydranet/experiments/exp_1762448567_7023_s96_f18_d3_m1_1_1_1
exp_1762448560_9364_s96_f16_d3_m2_3_3_4/model_info.json ADDED
@@ -0,0 +1,140 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "config": {
3
+ "input_size": 96,
4
+ "base_filters": 16,
5
+ "depth": 3,
6
+ "channel_multipliers": [
7
+ 2,
8
+ 3,
9
+ 3,
10
+ 4
11
+ ],
12
+ "n_channels": 8,
13
+ "n_classes": 3
14
+ },
15
+ "model_info": {
16
+ "depth": 3,
17
+ "channels_per_level": [
18
+ 32,
19
+ 48,
20
+ 48,
21
+ 64
22
+ ],
23
+ "channel_multipliers": [
24
+ 2,
25
+ 3,
26
+ 3,
27
+ 4
28
+ ],
29
+ "total_parameters": 194195,
30
+ "trainable_parameters": 194195,
31
+ "model_size_mb": 0.7407951354980469
32
+ },
33
+ "architecture_stats": {
34
+ "input_dimension": 73728,
35
+ "output_dimension": 27648,
36
+ "parameter_distribution": {
37
+ "encoder_params": 54224,
38
+ "decoder_params": 100000,
39
+ "bottleneck_params": 39552,
40
+ "other_params": 99,
41
+ "total_params": 193875,
42
+ "encoder_percentage": 27.968536428110895,
43
+ "decoder_percentage": 51.57962604771116,
44
+ "bottleneck_percentage": 20.40077369439072,
45
+ "other_percentage": 0.05106382978723404
46
+ },
47
+ "component_breakdown": {
48
+ "Encoder - encoders.0.channel_proj": 288,
49
+ "Encoder - encoders.0.convnext_block.dwconv": 1600,
50
+ "Encoder - encoders.0.convnext_block.norm": 64,
51
+ "Encoder - encoders.0.convnext_block.pwconv1": 4224,
52
+ "Encoder - encoders.0.convnext_block.pwconv2": 4128,
53
+ "Encoder - encoders.1.channel_proj": 1584,
54
+ "Encoder - encoders.1.convnext_block.dwconv": 2400,
55
+ "Encoder - encoders.1.convnext_block.norm": 96,
56
+ "Encoder - encoders.1.convnext_block.pwconv1": 9408,
57
+ "Encoder - encoders.1.convnext_block.pwconv2": 9264,
58
+ "Encoder - encoders.2.convnext_block.dwconv": 2400,
59
+ "Encoder - encoders.2.convnext_block.norm": 96,
60
+ "Encoder - encoders.2.convnext_block.pwconv1": 9408,
61
+ "Encoder - encoders.2.convnext_block.pwconv2": 9264,
62
+ "Bottleneck - bottleneck.channel_proj": 3136,
63
+ "Bottleneck - bottleneck.convnext_block.dwconv": 3200,
64
+ "Bottleneck - bottleneck.convnext_block.norm": 128,
65
+ "Bottleneck - bottleneck.convnext_block.pwconv1": 16640,
66
+ "Bottleneck - bottleneck.convnext_block.pwconv2": 16448,
67
+ "Decoder - upsamplers.0": 16448,
68
+ "Decoder - upsamplers.1": 9264,
69
+ "Decoder - upsamplers.2": 9264,
70
+ "Decoder - decoders.0.channel_proj": 5424,
71
+ "Decoder - decoders.0.convnext_block.dwconv": 2400,
72
+ "Decoder - decoders.0.convnext_block.norm": 96,
73
+ "Decoder - decoders.0.convnext_block.pwconv1": 9408,
74
+ "Decoder - decoders.0.convnext_block.pwconv2": 9264,
75
+ "Decoder - decoders.1.channel_proj": 4656,
76
+ "Decoder - decoders.1.convnext_block.dwconv": 2400,
77
+ "Decoder - decoders.1.convnext_block.norm": 96,
78
+ "Decoder - decoders.1.convnext_block.pwconv1": 9408,
79
+ "Decoder - decoders.1.convnext_block.pwconv2": 9264,
80
+ "Decoder - decoders.2.channel_proj": 2592,
81
+ "Decoder - decoders.2.convnext_block.dwconv": 1600,
82
+ "Decoder - decoders.2.convnext_block.norm": 64,
83
+ "Decoder - decoders.2.convnext_block.pwconv1": 4224,
84
+ "Decoder - decoders.2.convnext_block.pwconv2": 4128,
85
+ "Other - final_conv": 99
86
+ },
87
+ "latent_dimensions": {
88
+ "Level_0": {
89
+ "channels": 32,
90
+ "height": 96,
91
+ "width": 96,
92
+ "total_elements": 294912
93
+ },
94
+ "Level_1": {
95
+ "channels": 48,
96
+ "height": 48,
97
+ "width": 48,
98
+ "total_elements": 110592
99
+ },
100
+ "Level_2": {
101
+ "channels": 48,
102
+ "height": 24,
103
+ "width": 24,
104
+ "total_elements": 27648
105
+ },
106
+ "Level_3": {
107
+ "channels": 64,
108
+ "height": 12,
109
+ "width": 12,
110
+ "total_elements": 9216
111
+ }
112
+ },
113
+ "skip_dimensions": {
114
+ "Skip_0": {
115
+ "channels": 32,
116
+ "height": 96,
117
+ "width": 96,
118
+ "total_elements": 294912
119
+ },
120
+ "Skip_1": {
121
+ "channels": 48,
122
+ "height": 48,
123
+ "width": 48,
124
+ "total_elements": 110592
125
+ },
126
+ "Skip_2": {
127
+ "channels": 48,
128
+ "height": 24,
129
+ "width": 24,
130
+ "total_elements": 27648
131
+ }
132
+ },
133
+ "memory_analysis": {
134
+ "peak_memory_inference_mb": 1.96875,
135
+ "peak_memory_training_mb": 3.9375,
136
+ "peak_elements": 516096
137
+ }
138
+ },
139
+ "timestamp": "2025-11-06T18:02:40.283930"
140
+ }
exp_1762448560_9364_s96_f16_d3_m2_3_3_4/pipeline_results.json ADDED
@@ -0,0 +1,140 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "experiment_id": "exp_1762448560_4658_s96_f16_d3_m2_3_3_4",
3
+ "config": {
4
+ "input_size": 96,
5
+ "base_filters": 16,
6
+ "depth": 3,
7
+ "channel_multipliers": [
8
+ 2,
9
+ 3,
10
+ 3,
11
+ 4
12
+ ],
13
+ "n_channels": 8,
14
+ "n_classes": 3
15
+ },
16
+ "start_time": "2025-11-06T18:02:40.277074",
17
+ "success": true,
18
+ "timings": {
19
+ "onnx_conversion": 0.28563704597763717,
20
+ "openvino_conversion": 4.088858654955402,
21
+ "inference_total": 2.74204633012414,
22
+ "inference_actual": 0.132681,
23
+ "total": 7.129755483940244
24
+ },
25
+ "errors": [],
26
+ "architecture_stats": {
27
+ "input_dimension": 73728,
28
+ "output_dimension": 27648,
29
+ "parameter_distribution": {
30
+ "encoder_params": 54224,
31
+ "decoder_params": 100000,
32
+ "bottleneck_params": 39552,
33
+ "other_params": 99,
34
+ "total_params": 193875,
35
+ "encoder_percentage": 27.968536428110895,
36
+ "decoder_percentage": 51.57962604771116,
37
+ "bottleneck_percentage": 20.40077369439072,
38
+ "other_percentage": 0.05106382978723404
39
+ },
40
+ "component_breakdown": {
41
+ "Encoder - encoders.0.channel_proj": 288,
42
+ "Encoder - encoders.0.convnext_block.dwconv": 1600,
43
+ "Encoder - encoders.0.convnext_block.norm": 64,
44
+ "Encoder - encoders.0.convnext_block.pwconv1": 4224,
45
+ "Encoder - encoders.0.convnext_block.pwconv2": 4128,
46
+ "Encoder - encoders.1.channel_proj": 1584,
47
+ "Encoder - encoders.1.convnext_block.dwconv": 2400,
48
+ "Encoder - encoders.1.convnext_block.norm": 96,
49
+ "Encoder - encoders.1.convnext_block.pwconv1": 9408,
50
+ "Encoder - encoders.1.convnext_block.pwconv2": 9264,
51
+ "Encoder - encoders.2.convnext_block.dwconv": 2400,
52
+ "Encoder - encoders.2.convnext_block.norm": 96,
53
+ "Encoder - encoders.2.convnext_block.pwconv1": 9408,
54
+ "Encoder - encoders.2.convnext_block.pwconv2": 9264,
55
+ "Bottleneck - bottleneck.channel_proj": 3136,
56
+ "Bottleneck - bottleneck.convnext_block.dwconv": 3200,
57
+ "Bottleneck - bottleneck.convnext_block.norm": 128,
58
+ "Bottleneck - bottleneck.convnext_block.pwconv1": 16640,
59
+ "Bottleneck - bottleneck.convnext_block.pwconv2": 16448,
60
+ "Decoder - upsamplers.0": 16448,
61
+ "Decoder - upsamplers.1": 9264,
62
+ "Decoder - upsamplers.2": 9264,
63
+ "Decoder - decoders.0.channel_proj": 5424,
64
+ "Decoder - decoders.0.convnext_block.dwconv": 2400,
65
+ "Decoder - decoders.0.convnext_block.norm": 96,
66
+ "Decoder - decoders.0.convnext_block.pwconv1": 9408,
67
+ "Decoder - decoders.0.convnext_block.pwconv2": 9264,
68
+ "Decoder - decoders.1.channel_proj": 4656,
69
+ "Decoder - decoders.1.convnext_block.dwconv": 2400,
70
+ "Decoder - decoders.1.convnext_block.norm": 96,
71
+ "Decoder - decoders.1.convnext_block.pwconv1": 9408,
72
+ "Decoder - decoders.1.convnext_block.pwconv2": 9264,
73
+ "Decoder - decoders.2.channel_proj": 2592,
74
+ "Decoder - decoders.2.convnext_block.dwconv": 1600,
75
+ "Decoder - decoders.2.convnext_block.norm": 64,
76
+ "Decoder - decoders.2.convnext_block.pwconv1": 4224,
77
+ "Decoder - decoders.2.convnext_block.pwconv2": 4128,
78
+ "Other - final_conv": 99
79
+ },
80
+ "latent_dimensions": {
81
+ "Level_0": {
82
+ "channels": 32,
83
+ "height": 96,
84
+ "width": 96,
85
+ "total_elements": 294912
86
+ },
87
+ "Level_1": {
88
+ "channels": 48,
89
+ "height": 48,
90
+ "width": 48,
91
+ "total_elements": 110592
92
+ },
93
+ "Level_2": {
94
+ "channels": 48,
95
+ "height": 24,
96
+ "width": 24,
97
+ "total_elements": 27648
98
+ },
99
+ "Level_3": {
100
+ "channels": 64,
101
+ "height": 12,
102
+ "width": 12,
103
+ "total_elements": 9216
104
+ }
105
+ },
106
+ "skip_dimensions": {
107
+ "Skip_0": {
108
+ "channels": 32,
109
+ "height": 96,
110
+ "width": 96,
111
+ "total_elements": 294912
112
+ },
113
+ "Skip_1": {
114
+ "channels": 48,
115
+ "height": 48,
116
+ "width": 48,
117
+ "total_elements": 110592
118
+ },
119
+ "Skip_2": {
120
+ "channels": 48,
121
+ "height": 24,
122
+ "width": 24,
123
+ "total_elements": 27648
124
+ }
125
+ },
126
+ "memory_analysis": {
127
+ "peak_memory_inference_mb": 1.96875,
128
+ "peak_memory_training_mb": 3.9375,
129
+ "peak_elements": 516096
130
+ }
131
+ },
132
+ "inference_results": {
133
+ "success": true,
134
+ "total_time": 2.74204633012414,
135
+ "inference_time": 0.132681,
136
+ "stdout": "[setupvars.sh] OpenVINO environment initialized\nStarting Myriad inference...\nInput: /home/mount/experiments/exp_1762448560_9364_s96_f16_d3_m2_3_3_4/onnx/sample_input.npy\nModel XML: /home/mount/experiments/exp_1762448560_9364_s96_f16_d3_m2_3_3_4/openvino/model.xml\nModel BIN: /home/mount/experiments/exp_1762448560_9364_s96_f16_d3_m2_3_3_4/openvino/model.bin\n[OK] OpenVINO inference engine imported successfully\nDevice: MYRIAD\nModel XML: /home/mount/experiments/exp_1762448560_9364_s96_f16_d3_m2_3_3_4/openvino/model.xml\nModel BIN: /home/mount/experiments/exp_1762448560_9364_s96_f16_d3_m2_3_3_4/openvino/model.bin\nInput file: /home/mount/experiments/exp_1762448560_9364_s96_f16_d3_m2_3_3_4/onnx/sample_input.npy\nInitializing OpenVINO Runtime Core...\nAvailable devices:\n[E:] [BSL] found 0 ioexpander device\n ['CPU', 'GNA', 'MYRIAD']\nLoading network...\nInput blob: input\nInput shape: [1, 8, 96, 96]\nOutput blob: Conv_106\nOutput shape: [1, 3, 96, 96]\nLoading network to MYRIAD...\nLoading input data...\nInput data shape: (1, 8, 96, 96)\nInput data type: float32\nRunning inference...\n[OK] Inference completed!\nInference time: 0.132681 seconds\nOutput shape: (1, 3, 96, 96)\nOutput dtype: float32\nOutput range: [-0.395752, 0.620117]\nMyriad inference completed!\n",
137
+ "stderr": ""
138
+ },
139
+ "end_time": "2025-11-06T18:02:47.406819"
140
+ }
exp_1762448560_9364_s96_f16_d3_m2_3_3_4/run_inference.sh ADDED
@@ -0,0 +1,19 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #!/bin/bash
2
+ # Source OpenVINO environment
3
+ source /opt/intel/openvino/bin/setupvars.sh
4
+
5
+ PYPATH=/home/mount/scripts/simple_inference.py
6
+
7
+ echo "Starting Myriad inference..."
8
+ echo "Input: /home/mount/experiments/exp_1762448560_9364_s96_f16_d3_m2_3_3_4/onnx/sample_input.npy"
9
+ echo "Model XML: /home/mount/experiments/exp_1762448560_9364_s96_f16_d3_m2_3_3_4/openvino/model.xml"
10
+ echo "Model BIN: /home/mount/experiments/exp_1762448560_9364_s96_f16_d3_m2_3_3_4/openvino/model.bin"
11
+
12
+ python3 $PYPATH \
13
+ --input_filepath /home/mount/experiments/exp_1762448560_9364_s96_f16_d3_m2_3_3_4/onnx/sample_input.npy \
14
+ --device_name MYRIAD \
15
+ --model_bin /home/mount/experiments/exp_1762448560_9364_s96_f16_d3_m2_3_3_4/openvino/model.bin \
16
+ --model_xml /home/mount/experiments/exp_1762448560_9364_s96_f16_d3_m2_3_3_4/openvino/model.xml \
17
+ --output /home/mount/experiments/exp_1762448560_9364_s96_f16_d3_m2_3_3_4/inference/
18
+
19
+ echo "Myriad inference completed!"
exp_1762448567_7023_s96_f18_d3_m1_1_1_1/convert_openvino.sh ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #!/bin/bash
2
+ # Source OpenVINO environment
3
+ source /opt/intel/openvino/bin/setupvars.sh
4
+
5
+ # OpenVINO Model Optimizer script for ONNX model
6
+ PYPATH=/opt/intel/openvino_2020.3.194/deployment_tools/model_optimizer/mo.py
7
+ ONNX_MODEL=/home/mount/experiments/exp_1762448567_7023_s96_f18_d3_m1_1_1_1/onnx/model.onnx
8
+ OUTPUT_DIR=/home/mount/experiments/exp_1762448567_7023_s96_f18_d3_m1_1_1_1/openvino
9
+
10
+ echo "Converting ONNX model to OpenVINO IR format..."
11
+ echo "Input model: $ONNX_MODEL"
12
+ echo "Output directory: $OUTPUT_DIR"
13
+
14
+ # Create output directory if it doesn't exist
15
+ mkdir -p "$OUTPUT_DIR"
16
+
17
+ python3 $PYPATH \
18
+ --input_model "$ONNX_MODEL" \
19
+ --data_type FP16 \
20
+ --input_shape "[1,8,96,96]" \
21
+ --mean_values "[0,0,0,0,0,0,0,0]" \
22
+ --scale_values "[1,1,1,1,1,1,1,1]" \
23
+ --progress \
24
+ --stream_output \
25
+ --output_dir "$OUTPUT_DIR" \
26
+ --model_name model
27
+
28
+ echo "OpenVINO conversion completed!"
29
+ echo "Generated files in $OUTPUT_DIR:"
30
+ ls -la "$OUTPUT_DIR/"
exp_1762448567_7023_s96_f18_d3_m1_1_1_1/experiment.log ADDED
@@ -0,0 +1,98 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2025-11-06 18:02:47 - INFO - === Model Creation Phase ===
2
+ 2025-11-06 18:02:47 - INFO - Creating UNet model with config: {'input_size': 96, 'base_filters': 18, 'depth': 3, 'channel_multipliers': [1, 1, 1, 1], 'n_channels': 8, 'n_classes': 3}
3
+ 2025-11-06 18:02:47 - INFO - Model created successfully:
4
+ 2025-11-06 18:02:47 - INFO - Depth: 3
5
+ 2025-11-06 18:02:47 - INFO - Channels per level: [18, 18, 18, 18]
6
+ 2025-11-06 18:02:47 - INFO - Total parameters: 31,611
7
+ 2025-11-06 18:02:47 - INFO - Parameter memory: 0.12 MB
8
+ 2025-11-06 18:02:47 - INFO - === Detailed Architecture Analysis ===
9
+ 2025-11-06 18:02:47 - INFO - Input vector dimension: 73,728 (8 × 96 × 96)
10
+ 2025-11-06 18:02:47 - INFO - Component-wise parameter breakdown:
11
+ 2025-11-06 18:02:47 - INFO - Encoder - encoders.0.channel_proj: 162 parameters
12
+ 2025-11-06 18:02:47 - INFO - Encoder - encoders.0.convnext_block.dwconv: 900 parameters
13
+ 2025-11-06 18:02:47 - INFO - Encoder - encoders.0.convnext_block.norm: 36 parameters
14
+ 2025-11-06 18:02:47 - INFO - Encoder - encoders.0.convnext_block.pwconv1: 1,368 parameters
15
+ 2025-11-06 18:02:47 - INFO - Encoder - encoders.0.convnext_block.pwconv2: 1,314 parameters
16
+ 2025-11-06 18:02:47 - INFO - Encoder - encoders.1.convnext_block.dwconv: 900 parameters
17
+ 2025-11-06 18:02:47 - INFO - Encoder - encoders.1.convnext_block.norm: 36 parameters
18
+ 2025-11-06 18:02:47 - INFO - Encoder - encoders.1.convnext_block.pwconv1: 1,368 parameters
19
+ 2025-11-06 18:02:47 - INFO - Encoder - encoders.1.convnext_block.pwconv2: 1,314 parameters
20
+ 2025-11-06 18:02:47 - INFO - Encoder - encoders.2.convnext_block.dwconv: 900 parameters
21
+ 2025-11-06 18:02:47 - INFO - Encoder - encoders.2.convnext_block.norm: 36 parameters
22
+ 2025-11-06 18:02:47 - INFO - Encoder - encoders.2.convnext_block.pwconv1: 1,368 parameters
23
+ 2025-11-06 18:02:47 - INFO - Encoder - encoders.2.convnext_block.pwconv2: 1,314 parameters
24
+ 2025-11-06 18:02:47 - INFO - Bottleneck - bottleneck.convnext_block.dwconv: 900 parameters
25
+ 2025-11-06 18:02:47 - INFO - Bottleneck - bottleneck.convnext_block.norm: 36 parameters
26
+ 2025-11-06 18:02:47 - INFO - Bottleneck - bottleneck.convnext_block.pwconv1: 1,368 parameters
27
+ 2025-11-06 18:02:47 - INFO - Bottleneck - bottleneck.convnext_block.pwconv2: 1,314 parameters
28
+ 2025-11-06 18:02:47 - INFO - Decoder - upsamplers.0: 1,314 parameters
29
+ 2025-11-06 18:02:47 - INFO - Decoder - upsamplers.1: 1,314 parameters
30
+ 2025-11-06 18:02:47 - INFO - Decoder - upsamplers.2: 1,314 parameters
31
+ 2025-11-06 18:02:47 - INFO - Decoder - decoders.0.channel_proj: 666 parameters
32
+ 2025-11-06 18:02:47 - INFO - Decoder - decoders.0.convnext_block.dwconv: 900 parameters
33
+ 2025-11-06 18:02:47 - INFO - Decoder - decoders.0.convnext_block.norm: 36 parameters
34
+ 2025-11-06 18:02:47 - INFO - Decoder - decoders.0.convnext_block.pwconv1: 1,368 parameters
35
+ 2025-11-06 18:02:47 - INFO - Decoder - decoders.0.convnext_block.pwconv2: 1,314 parameters
36
+ 2025-11-06 18:02:47 - INFO - Decoder - decoders.1.channel_proj: 666 parameters
37
+ 2025-11-06 18:02:47 - INFO - Decoder - decoders.1.convnext_block.dwconv: 900 parameters
38
+ 2025-11-06 18:02:47 - INFO - Decoder - decoders.1.convnext_block.norm: 36 parameters
39
+ 2025-11-06 18:02:47 - INFO - Decoder - decoders.1.convnext_block.pwconv1: 1,368 parameters
40
+ 2025-11-06 18:02:47 - INFO - Decoder - decoders.1.convnext_block.pwconv2: 1,314 parameters
41
+ 2025-11-06 18:02:47 - INFO - Decoder - decoders.2.channel_proj: 666 parameters
42
+ 2025-11-06 18:02:47 - INFO - Decoder - decoders.2.convnext_block.dwconv: 900 parameters
43
+ 2025-11-06 18:02:47 - INFO - Decoder - decoders.2.convnext_block.norm: 36 parameters
44
+ 2025-11-06 18:02:47 - INFO - Decoder - decoders.2.convnext_block.pwconv1: 1,368 parameters
45
+ 2025-11-06 18:02:47 - INFO - Decoder - decoders.2.convnext_block.pwconv2: 1,314 parameters
46
+ 2025-11-06 18:02:47 - INFO - Other - final_conv: 57 parameters
47
+ 2025-11-06 18:02:47 - INFO - Parameter distribution summary:
48
+ 2025-11-06 18:02:47 - INFO - Encoder parameters: 11,016 (35.0%)
49
+ 2025-11-06 18:02:47 - INFO - Decoder parameters: 16,794 (53.3%)
50
+ 2025-11-06 18:02:47 - INFO - Bottleneck parameters: 3,618 (11.5%)
51
+ 2025-11-06 18:02:47 - INFO - Other parameters: 57 (0.2%)
52
+ 2025-11-06 18:02:47 - INFO - Latent space dimensions (feature maps at each level):
53
+ 2025-11-06 18:02:47 - INFO - Level 0: 18 × 96 × 96 = 165,888 elements
54
+ 2025-11-06 18:02:47 - INFO - Level 1: 18 × 48 × 48 = 41,472 elements
55
+ 2025-11-06 18:02:47 - INFO - Level 2: 18 × 24 × 24 = 10,368 elements
56
+ 2025-11-06 18:02:47 - INFO - Level 3: 18 × 12 × 12 = 2,592 elements
57
+ 2025-11-06 18:02:47 - INFO - Skip connection dimensions:
58
+ 2025-11-06 18:02:47 - INFO - Skip 0: 18 × 96 × 96 = 165,888 elements
59
+ 2025-11-06 18:02:47 - INFO - Skip 1: 18 × 48 × 48 = 41,472 elements
60
+ 2025-11-06 18:02:47 - INFO - Skip 2: 18 × 24 × 24 = 10,368 elements
61
+ 2025-11-06 18:02:47 - INFO - Memory analysis:
62
+ 2025-11-06 18:02:47 - INFO - Peak feature map memory (inference): 1.12 MB
63
+ 2025-11-06 18:02:47 - INFO - Peak feature map memory (training): 2.24 MB (with gradients)
64
+ 2025-11-06 18:02:47 - INFO - Output vector dimension: 27,648 (3 × 96 × 96)
65
+ 2025-11-06 18:02:47 - INFO - PyTorch model saved to: /home/philab/Desktop/hydranet/experiments/exp_1762448567_7023_s96_f18_d3_m1_1_1_1/pytorch/model.pt
66
+ 2025-11-06 18:02:47 - INFO - === ONNX Conversion Phase ===
67
+ 2025-11-06 18:02:47 - INFO - === Model Export Diagnostics ===
68
+ 2025-11-06 18:02:47 - INFO - PyTorch version: 1.9.0+cu102
69
+ 2025-11-06 18:02:47 - INFO - Model parameters: 31,611
70
+ 2025-11-06 18:02:47 - INFO - Model memory: 0.12 MB
71
+ 2025-11-06 18:02:47 - INFO - Starting ONNX export with opset version 11
72
+ 2025-11-06 18:02:47 - INFO - Model input shape: torch.Size([1, 8, 96, 96])
73
+ 2025-11-06 18:02:47 - INFO - Model input dtype: torch.float32
74
+ 2025-11-06 18:02:47 - INFO - Forward pass successful. Output shape: torch.Size([1, 3, 96, 96])
75
+ 2025-11-06 18:02:47 - INFO - Output dtype: torch.float32
76
+ 2025-11-06 18:02:47 - INFO - Output value range: [-1.0509, 0.4186]
77
+ 2025-11-06 18:02:47 - INFO - Model successfully exported to /home/philab/Desktop/hydranet/experiments/exp_1762448567_7023_s96_f18_d3_m1_1_1_1/onnx/model.onnx
78
+ 2025-11-06 18:02:47 - INFO - ONNX model size: 0.13 MB
79
+ 2025-11-06 18:02:47 - INFO - Saved dummy input with shape (1, 8, 96, 96) to /home/philab/Desktop/hydranet/experiments/exp_1762448567_7023_s96_f18_d3_m1_1_1_1/onnx/sample_input.npy
80
+ 2025-11-06 18:02:47 - INFO - Input data type: float32
81
+ 2025-11-06 18:02:47 - INFO - Input value range: [-4.3423, 4.4816]
82
+ 2025-11-06 18:02:47 - INFO - === OpenVINO Conversion Phase ===
83
+ 2025-11-06 18:02:47 - INFO - Starting OpenVINO conversion in Docker container...
84
+ 2025-11-06 18:02:51 - INFO - OpenVINO conversion completed in 3.97 seconds
85
+ 2025-11-06 18:02:51 - INFO - OpenVINO model files created:
86
+ 2025-11-06 18:02:51 - INFO - XML file: /home/philab/Desktop/hydranet/experiments/exp_1762448567_7023_s96_f18_d3_m1_1_1_1/openvino/model.xml (0.08 MB)
87
+ 2025-11-06 18:02:51 - INFO - BIN file: /home/philab/Desktop/hydranet/experiments/exp_1762448567_7023_s96_f18_d3_m1_1_1_1/openvino/model.bin (0.06 MB)
88
+ 2025-11-06 18:02:51 - INFO - === Myriad Inference Phase ===
89
+ 2025-11-06 18:02:51 - INFO - Starting Myriad inference in Docker container...
90
+ 2025-11-06 18:02:54 - INFO - Myriad inference completed in 2.67 seconds
91
+ 2025-11-06 18:02:54 - INFO - Actual inference time: 0.118388 seconds
92
+ 2025-11-06 18:02:54 - INFO - ✅ Complete pipeline executed successfully!
93
+ 2025-11-06 18:02:54 - INFO - ✅ Experiment 12 completed successfully
94
+ 2025-11-06 18:02:54 - INFO - Inference time: 0.118388s
95
+ 2025-11-06 18:02:54 - INFO -
96
+ === Experiment 13/2475 ===
97
+ 2025-11-06 18:02:54 - INFO - Experiment ID: exp_1762448574_8263_s96_f18_d3_m2_2_2_2
98
+ 2025-11-06 18:02:54 - INFO - Experiment directory: /home/philab/Desktop/hydranet/experiments/exp_1762448574_8263_s96_f18_d3_m2_2_2_2
exp_1762448567_7023_s96_f18_d3_m1_1_1_1/model_info.json ADDED
@@ -0,0 +1,138 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "config": {
3
+ "input_size": 96,
4
+ "base_filters": 18,
5
+ "depth": 3,
6
+ "channel_multipliers": [
7
+ 1,
8
+ 1,
9
+ 1,
10
+ 1
11
+ ],
12
+ "n_channels": 8,
13
+ "n_classes": 3
14
+ },
15
+ "model_info": {
16
+ "depth": 3,
17
+ "channels_per_level": [
18
+ 18,
19
+ 18,
20
+ 18,
21
+ 18
22
+ ],
23
+ "channel_multipliers": [
24
+ 1,
25
+ 1,
26
+ 1,
27
+ 1
28
+ ],
29
+ "total_parameters": 31611,
30
+ "trainable_parameters": 31611,
31
+ "model_size_mb": 0.12058639526367188
32
+ },
33
+ "architecture_stats": {
34
+ "input_dimension": 73728,
35
+ "output_dimension": 27648,
36
+ "parameter_distribution": {
37
+ "encoder_params": 11016,
38
+ "decoder_params": 16794,
39
+ "bottleneck_params": 3618,
40
+ "other_params": 57,
41
+ "total_params": 31485,
42
+ "encoder_percentage": 34.98808956646022,
43
+ "decoder_percentage": 53.33968556455455,
44
+ "bottleneck_percentage": 11.491186279180562,
45
+ "other_percentage": 0.1810385898046689
46
+ },
47
+ "component_breakdown": {
48
+ "Encoder - encoders.0.channel_proj": 162,
49
+ "Encoder - encoders.0.convnext_block.dwconv": 900,
50
+ "Encoder - encoders.0.convnext_block.norm": 36,
51
+ "Encoder - encoders.0.convnext_block.pwconv1": 1368,
52
+ "Encoder - encoders.0.convnext_block.pwconv2": 1314,
53
+ "Encoder - encoders.1.convnext_block.dwconv": 900,
54
+ "Encoder - encoders.1.convnext_block.norm": 36,
55
+ "Encoder - encoders.1.convnext_block.pwconv1": 1368,
56
+ "Encoder - encoders.1.convnext_block.pwconv2": 1314,
57
+ "Encoder - encoders.2.convnext_block.dwconv": 900,
58
+ "Encoder - encoders.2.convnext_block.norm": 36,
59
+ "Encoder - encoders.2.convnext_block.pwconv1": 1368,
60
+ "Encoder - encoders.2.convnext_block.pwconv2": 1314,
61
+ "Bottleneck - bottleneck.convnext_block.dwconv": 900,
62
+ "Bottleneck - bottleneck.convnext_block.norm": 36,
63
+ "Bottleneck - bottleneck.convnext_block.pwconv1": 1368,
64
+ "Bottleneck - bottleneck.convnext_block.pwconv2": 1314,
65
+ "Decoder - upsamplers.0": 1314,
66
+ "Decoder - upsamplers.1": 1314,
67
+ "Decoder - upsamplers.2": 1314,
68
+ "Decoder - decoders.0.channel_proj": 666,
69
+ "Decoder - decoders.0.convnext_block.dwconv": 900,
70
+ "Decoder - decoders.0.convnext_block.norm": 36,
71
+ "Decoder - decoders.0.convnext_block.pwconv1": 1368,
72
+ "Decoder - decoders.0.convnext_block.pwconv2": 1314,
73
+ "Decoder - decoders.1.channel_proj": 666,
74
+ "Decoder - decoders.1.convnext_block.dwconv": 900,
75
+ "Decoder - decoders.1.convnext_block.norm": 36,
76
+ "Decoder - decoders.1.convnext_block.pwconv1": 1368,
77
+ "Decoder - decoders.1.convnext_block.pwconv2": 1314,
78
+ "Decoder - decoders.2.channel_proj": 666,
79
+ "Decoder - decoders.2.convnext_block.dwconv": 900,
80
+ "Decoder - decoders.2.convnext_block.norm": 36,
81
+ "Decoder - decoders.2.convnext_block.pwconv1": 1368,
82
+ "Decoder - decoders.2.convnext_block.pwconv2": 1314,
83
+ "Other - final_conv": 57
84
+ },
85
+ "latent_dimensions": {
86
+ "Level_0": {
87
+ "channels": 18,
88
+ "height": 96,
89
+ "width": 96,
90
+ "total_elements": 165888
91
+ },
92
+ "Level_1": {
93
+ "channels": 18,
94
+ "height": 48,
95
+ "width": 48,
96
+ "total_elements": 41472
97
+ },
98
+ "Level_2": {
99
+ "channels": 18,
100
+ "height": 24,
101
+ "width": 24,
102
+ "total_elements": 10368
103
+ },
104
+ "Level_3": {
105
+ "channels": 18,
106
+ "height": 12,
107
+ "width": 12,
108
+ "total_elements": 2592
109
+ }
110
+ },
111
+ "skip_dimensions": {
112
+ "Skip_0": {
113
+ "channels": 18,
114
+ "height": 96,
115
+ "width": 96,
116
+ "total_elements": 165888
117
+ },
118
+ "Skip_1": {
119
+ "channels": 18,
120
+ "height": 48,
121
+ "width": 48,
122
+ "total_elements": 41472
123
+ },
124
+ "Skip_2": {
125
+ "channels": 18,
126
+ "height": 24,
127
+ "width": 24,
128
+ "total_elements": 10368
129
+ }
130
+ },
131
+ "memory_analysis": {
132
+ "peak_memory_inference_mb": 1.1217041015625,
133
+ "peak_memory_training_mb": 2.243408203125,
134
+ "peak_elements": 294048
135
+ }
136
+ },
137
+ "timestamp": "2025-11-06T18:02:47.415174"
138
+ }
exp_1762448567_7023_s96_f18_d3_m1_1_1_1/pipeline_results.json ADDED
@@ -0,0 +1,138 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "experiment_id": "exp_1762448567_2535_s96_f18_d3_m1_1_1_1",
3
+ "config": {
4
+ "input_size": 96,
5
+ "base_filters": 18,
6
+ "depth": 3,
7
+ "channel_multipliers": [
8
+ 1,
9
+ 1,
10
+ 1,
11
+ 1
12
+ ],
13
+ "n_channels": 8,
14
+ "n_classes": 3
15
+ },
16
+ "start_time": "2025-11-06T18:02:47.409086",
17
+ "success": true,
18
+ "timings": {
19
+ "onnx_conversion": 0.25858570099808276,
20
+ "openvino_conversion": 3.9692281188908964,
21
+ "inference_total": 2.66881703492254,
22
+ "inference_actual": 0.118388,
23
+ "total": 6.908043346134946
24
+ },
25
+ "errors": [],
26
+ "architecture_stats": {
27
+ "input_dimension": 73728,
28
+ "output_dimension": 27648,
29
+ "parameter_distribution": {
30
+ "encoder_params": 11016,
31
+ "decoder_params": 16794,
32
+ "bottleneck_params": 3618,
33
+ "other_params": 57,
34
+ "total_params": 31485,
35
+ "encoder_percentage": 34.98808956646022,
36
+ "decoder_percentage": 53.33968556455455,
37
+ "bottleneck_percentage": 11.491186279180562,
38
+ "other_percentage": 0.1810385898046689
39
+ },
40
+ "component_breakdown": {
41
+ "Encoder - encoders.0.channel_proj": 162,
42
+ "Encoder - encoders.0.convnext_block.dwconv": 900,
43
+ "Encoder - encoders.0.convnext_block.norm": 36,
44
+ "Encoder - encoders.0.convnext_block.pwconv1": 1368,
45
+ "Encoder - encoders.0.convnext_block.pwconv2": 1314,
46
+ "Encoder - encoders.1.convnext_block.dwconv": 900,
47
+ "Encoder - encoders.1.convnext_block.norm": 36,
48
+ "Encoder - encoders.1.convnext_block.pwconv1": 1368,
49
+ "Encoder - encoders.1.convnext_block.pwconv2": 1314,
50
+ "Encoder - encoders.2.convnext_block.dwconv": 900,
51
+ "Encoder - encoders.2.convnext_block.norm": 36,
52
+ "Encoder - encoders.2.convnext_block.pwconv1": 1368,
53
+ "Encoder - encoders.2.convnext_block.pwconv2": 1314,
54
+ "Bottleneck - bottleneck.convnext_block.dwconv": 900,
55
+ "Bottleneck - bottleneck.convnext_block.norm": 36,
56
+ "Bottleneck - bottleneck.convnext_block.pwconv1": 1368,
57
+ "Bottleneck - bottleneck.convnext_block.pwconv2": 1314,
58
+ "Decoder - upsamplers.0": 1314,
59
+ "Decoder - upsamplers.1": 1314,
60
+ "Decoder - upsamplers.2": 1314,
61
+ "Decoder - decoders.0.channel_proj": 666,
62
+ "Decoder - decoders.0.convnext_block.dwconv": 900,
63
+ "Decoder - decoders.0.convnext_block.norm": 36,
64
+ "Decoder - decoders.0.convnext_block.pwconv1": 1368,
65
+ "Decoder - decoders.0.convnext_block.pwconv2": 1314,
66
+ "Decoder - decoders.1.channel_proj": 666,
67
+ "Decoder - decoders.1.convnext_block.dwconv": 900,
68
+ "Decoder - decoders.1.convnext_block.norm": 36,
69
+ "Decoder - decoders.1.convnext_block.pwconv1": 1368,
70
+ "Decoder - decoders.1.convnext_block.pwconv2": 1314,
71
+ "Decoder - decoders.2.channel_proj": 666,
72
+ "Decoder - decoders.2.convnext_block.dwconv": 900,
73
+ "Decoder - decoders.2.convnext_block.norm": 36,
74
+ "Decoder - decoders.2.convnext_block.pwconv1": 1368,
75
+ "Decoder - decoders.2.convnext_block.pwconv2": 1314,
76
+ "Other - final_conv": 57
77
+ },
78
+ "latent_dimensions": {
79
+ "Level_0": {
80
+ "channels": 18,
81
+ "height": 96,
82
+ "width": 96,
83
+ "total_elements": 165888
84
+ },
85
+ "Level_1": {
86
+ "channels": 18,
87
+ "height": 48,
88
+ "width": 48,
89
+ "total_elements": 41472
90
+ },
91
+ "Level_2": {
92
+ "channels": 18,
93
+ "height": 24,
94
+ "width": 24,
95
+ "total_elements": 10368
96
+ },
97
+ "Level_3": {
98
+ "channels": 18,
99
+ "height": 12,
100
+ "width": 12,
101
+ "total_elements": 2592
102
+ }
103
+ },
104
+ "skip_dimensions": {
105
+ "Skip_0": {
106
+ "channels": 18,
107
+ "height": 96,
108
+ "width": 96,
109
+ "total_elements": 165888
110
+ },
111
+ "Skip_1": {
112
+ "channels": 18,
113
+ "height": 48,
114
+ "width": 48,
115
+ "total_elements": 41472
116
+ },
117
+ "Skip_2": {
118
+ "channels": 18,
119
+ "height": 24,
120
+ "width": 24,
121
+ "total_elements": 10368
122
+ }
123
+ },
124
+ "memory_analysis": {
125
+ "peak_memory_inference_mb": 1.1217041015625,
126
+ "peak_memory_training_mb": 2.243408203125,
127
+ "peak_elements": 294048
128
+ }
129
+ },
130
+ "inference_results": {
131
+ "success": true,
132
+ "total_time": 2.66881703492254,
133
+ "inference_time": 0.118388,
134
+ "stdout": "[setupvars.sh] OpenVINO environment initialized\nStarting Myriad inference...\nInput: /home/mount/experiments/exp_1762448567_7023_s96_f18_d3_m1_1_1_1/onnx/sample_input.npy\nModel XML: /home/mount/experiments/exp_1762448567_7023_s96_f18_d3_m1_1_1_1/openvino/model.xml\nModel BIN: /home/mount/experiments/exp_1762448567_7023_s96_f18_d3_m1_1_1_1/openvino/model.bin\n[OK] OpenVINO inference engine imported successfully\nDevice: MYRIAD\nModel XML: /home/mount/experiments/exp_1762448567_7023_s96_f18_d3_m1_1_1_1/openvino/model.xml\nModel BIN: /home/mount/experiments/exp_1762448567_7023_s96_f18_d3_m1_1_1_1/openvino/model.bin\nInput file: /home/mount/experiments/exp_1762448567_7023_s96_f18_d3_m1_1_1_1/onnx/sample_input.npy\nInitializing OpenVINO Runtime Core...\nAvailable devices:\n[E:] [BSL] found 0 ioexpander device\n ['CPU', 'GNA', 'MYRIAD']\nLoading network...\nInput blob: input\nInput shape: [1, 8, 96, 96]\nOutput blob: Conv_104\nOutput shape: [1, 3, 96, 96]\nLoading network to MYRIAD...\nLoading input data...\nInput data shape: (1, 8, 96, 96)\nInput data type: float32\nRunning inference...\n[OK] Inference completed!\nInference time: 0.118388 seconds\nOutput shape: (1, 3, 96, 96)\nOutput dtype: float32\nOutput range: [-1.050781, 0.418945]\nMyriad inference completed!\n",
135
+ "stderr": ""
136
+ },
137
+ "end_time": "2025-11-06T18:02:54.317117"
138
+ }
exp_1762448567_7023_s96_f18_d3_m1_1_1_1/run_inference.sh ADDED
@@ -0,0 +1,19 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #!/bin/bash
2
+ # Source OpenVINO environment
3
+ source /opt/intel/openvino/bin/setupvars.sh
4
+
5
+ PYPATH=/home/mount/scripts/simple_inference.py
6
+
7
+ echo "Starting Myriad inference..."
8
+ echo "Input: /home/mount/experiments/exp_1762448567_7023_s96_f18_d3_m1_1_1_1/onnx/sample_input.npy"
9
+ echo "Model XML: /home/mount/experiments/exp_1762448567_7023_s96_f18_d3_m1_1_1_1/openvino/model.xml"
10
+ echo "Model BIN: /home/mount/experiments/exp_1762448567_7023_s96_f18_d3_m1_1_1_1/openvino/model.bin"
11
+
12
+ python3 $PYPATH \
13
+ --input_filepath /home/mount/experiments/exp_1762448567_7023_s96_f18_d3_m1_1_1_1/onnx/sample_input.npy \
14
+ --device_name MYRIAD \
15
+ --model_bin /home/mount/experiments/exp_1762448567_7023_s96_f18_d3_m1_1_1_1/openvino/model.bin \
16
+ --model_xml /home/mount/experiments/exp_1762448567_7023_s96_f18_d3_m1_1_1_1/openvino/model.xml \
17
+ --output /home/mount/experiments/exp_1762448567_7023_s96_f18_d3_m1_1_1_1/inference/
18
+
19
+ echo "Myriad inference completed!"
exp_1762448652_9207_s96_f20_d3_m2_2_2_2/convert_openvino.sh ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #!/bin/bash
2
+ # Source OpenVINO environment
3
+ source /opt/intel/openvino/bin/setupvars.sh
4
+
5
+ # OpenVINO Model Optimizer script for ONNX model
6
+ PYPATH=/opt/intel/openvino_2020.3.194/deployment_tools/model_optimizer/mo.py
7
+ ONNX_MODEL=/home/mount/experiments/exp_1762448652_9207_s96_f20_d3_m2_2_2_2/onnx/model.onnx
8
+ OUTPUT_DIR=/home/mount/experiments/exp_1762448652_9207_s96_f20_d3_m2_2_2_2/openvino
9
+
10
+ echo "Converting ONNX model to OpenVINO IR format..."
11
+ echo "Input model: $ONNX_MODEL"
12
+ echo "Output directory: $OUTPUT_DIR"
13
+
14
+ # Create output directory if it doesn't exist
15
+ mkdir -p "$OUTPUT_DIR"
16
+
17
+ python3 $PYPATH \
18
+ --input_model "$ONNX_MODEL" \
19
+ --data_type FP16 \
20
+ --input_shape "[1,8,96,96]" \
21
+ --mean_values "[0,0,0,0,0,0,0,0]" \
22
+ --scale_values "[1,1,1,1,1,1,1,1]" \
23
+ --progress \
24
+ --stream_output \
25
+ --output_dir "$OUTPUT_DIR" \
26
+ --model_name model
27
+
28
+ echo "OpenVINO conversion completed!"
29
+ echo "Generated files in $OUTPUT_DIR:"
30
+ ls -la "$OUTPUT_DIR/"
exp_1762448652_9207_s96_f20_d3_m2_2_2_2/experiment.log ADDED
@@ -0,0 +1,98 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2025-11-06 18:04:12 - INFO - === Model Creation Phase ===
2
+ 2025-11-06 18:04:12 - INFO - Creating UNet model with config: {'input_size': 96, 'base_filters': 20, 'depth': 3, 'channel_multipliers': [2, 2, 2, 2], 'n_channels': 8, 'n_classes': 3}
3
+ 2025-11-06 18:04:12 - INFO - Model created successfully:
4
+ 2025-11-06 18:04:12 - INFO - Depth: 3
5
+ 2025-11-06 18:04:12 - INFO - Channels per level: [40, 40, 40, 40]
6
+ 2025-11-06 18:04:12 - INFO - Total parameters: 135,363
7
+ 2025-11-06 18:04:12 - INFO - Parameter memory: 0.52 MB
8
+ 2025-11-06 18:04:12 - INFO - === Detailed Architecture Analysis ===
9
+ 2025-11-06 18:04:12 - INFO - Input vector dimension: 73,728 (8 × 96 × 96)
10
+ 2025-11-06 18:04:12 - INFO - Component-wise parameter breakdown:
11
+ 2025-11-06 18:04:12 - INFO - Encoder - encoders.0.channel_proj: 360 parameters
12
+ 2025-11-06 18:04:12 - INFO - Encoder - encoders.0.convnext_block.dwconv: 2,000 parameters
13
+ 2025-11-06 18:04:12 - INFO - Encoder - encoders.0.convnext_block.norm: 80 parameters
14
+ 2025-11-06 18:04:12 - INFO - Encoder - encoders.0.convnext_block.pwconv1: 6,560 parameters
15
+ 2025-11-06 18:04:12 - INFO - Encoder - encoders.0.convnext_block.pwconv2: 6,440 parameters
16
+ 2025-11-06 18:04:12 - INFO - Encoder - encoders.1.convnext_block.dwconv: 2,000 parameters
17
+ 2025-11-06 18:04:12 - INFO - Encoder - encoders.1.convnext_block.norm: 80 parameters
18
+ 2025-11-06 18:04:12 - INFO - Encoder - encoders.1.convnext_block.pwconv1: 6,560 parameters
19
+ 2025-11-06 18:04:12 - INFO - Encoder - encoders.1.convnext_block.pwconv2: 6,440 parameters
20
+ 2025-11-06 18:04:12 - INFO - Encoder - encoders.2.convnext_block.dwconv: 2,000 parameters
21
+ 2025-11-06 18:04:12 - INFO - Encoder - encoders.2.convnext_block.norm: 80 parameters
22
+ 2025-11-06 18:04:12 - INFO - Encoder - encoders.2.convnext_block.pwconv1: 6,560 parameters
23
+ 2025-11-06 18:04:12 - INFO - Encoder - encoders.2.convnext_block.pwconv2: 6,440 parameters
24
+ 2025-11-06 18:04:12 - INFO - Bottleneck - bottleneck.convnext_block.dwconv: 2,000 parameters
25
+ 2025-11-06 18:04:12 - INFO - Bottleneck - bottleneck.convnext_block.norm: 80 parameters
26
+ 2025-11-06 18:04:12 - INFO - Bottleneck - bottleneck.convnext_block.pwconv1: 6,560 parameters
27
+ 2025-11-06 18:04:12 - INFO - Bottleneck - bottleneck.convnext_block.pwconv2: 6,440 parameters
28
+ 2025-11-06 18:04:12 - INFO - Decoder - upsamplers.0: 6,440 parameters
29
+ 2025-11-06 18:04:12 - INFO - Decoder - upsamplers.1: 6,440 parameters
30
+ 2025-11-06 18:04:12 - INFO - Decoder - upsamplers.2: 6,440 parameters
31
+ 2025-11-06 18:04:12 - INFO - Decoder - decoders.0.channel_proj: 3,240 parameters
32
+ 2025-11-06 18:04:12 - INFO - Decoder - decoders.0.convnext_block.dwconv: 2,000 parameters
33
+ 2025-11-06 18:04:12 - INFO - Decoder - decoders.0.convnext_block.norm: 80 parameters
34
+ 2025-11-06 18:04:12 - INFO - Decoder - decoders.0.convnext_block.pwconv1: 6,560 parameters
35
+ 2025-11-06 18:04:12 - INFO - Decoder - decoders.0.convnext_block.pwconv2: 6,440 parameters
36
+ 2025-11-06 18:04:12 - INFO - Decoder - decoders.1.channel_proj: 3,240 parameters
37
+ 2025-11-06 18:04:12 - INFO - Decoder - decoders.1.convnext_block.dwconv: 2,000 parameters
38
+ 2025-11-06 18:04:12 - INFO - Decoder - decoders.1.convnext_block.norm: 80 parameters
39
+ 2025-11-06 18:04:12 - INFO - Decoder - decoders.1.convnext_block.pwconv1: 6,560 parameters
40
+ 2025-11-06 18:04:12 - INFO - Decoder - decoders.1.convnext_block.pwconv2: 6,440 parameters
41
+ 2025-11-06 18:04:12 - INFO - Decoder - decoders.2.channel_proj: 3,240 parameters
42
+ 2025-11-06 18:04:12 - INFO - Decoder - decoders.2.convnext_block.dwconv: 2,000 parameters
43
+ 2025-11-06 18:04:12 - INFO - Decoder - decoders.2.convnext_block.norm: 80 parameters
44
+ 2025-11-06 18:04:12 - INFO - Decoder - decoders.2.convnext_block.pwconv1: 6,560 parameters
45
+ 2025-11-06 18:04:12 - INFO - Decoder - decoders.2.convnext_block.pwconv2: 6,440 parameters
46
+ 2025-11-06 18:04:12 - INFO - Other - final_conv: 123 parameters
47
+ 2025-11-06 18:04:12 - INFO - Parameter distribution summary:
48
+ 2025-11-06 18:04:12 - INFO - Encoder parameters: 45,600 (33.8%)
49
+ 2025-11-06 18:04:12 - INFO - Decoder parameters: 74,280 (55.0%)
50
+ 2025-11-06 18:04:12 - INFO - Bottleneck parameters: 15,080 (11.2%)
51
+ 2025-11-06 18:04:12 - INFO - Other parameters: 123 (0.1%)
52
+ 2025-11-06 18:04:12 - INFO - Latent space dimensions (feature maps at each level):
53
+ 2025-11-06 18:04:12 - INFO - Level 0: 40 × 96 × 96 = 368,640 elements
54
+ 2025-11-06 18:04:12 - INFO - Level 1: 40 × 48 × 48 = 92,160 elements
55
+ 2025-11-06 18:04:12 - INFO - Level 2: 40 × 24 × 24 = 23,040 elements
56
+ 2025-11-06 18:04:12 - INFO - Level 3: 40 × 12 × 12 = 5,760 elements
57
+ 2025-11-06 18:04:12 - INFO - Skip connection dimensions:
58
+ 2025-11-06 18:04:12 - INFO - Skip 0: 40 × 96 × 96 = 368,640 elements
59
+ 2025-11-06 18:04:12 - INFO - Skip 1: 40 × 48 × 48 = 92,160 elements
60
+ 2025-11-06 18:04:12 - INFO - Skip 2: 40 × 24 × 24 = 23,040 elements
61
+ 2025-11-06 18:04:12 - INFO - Memory analysis:
62
+ 2025-11-06 18:04:12 - INFO - Peak feature map memory (inference): 2.15 MB
63
+ 2025-11-06 18:04:12 - INFO - Peak feature map memory (training): 4.30 MB (with gradients)
64
+ 2025-11-06 18:04:12 - INFO - Output vector dimension: 27,648 (3 × 96 × 96)
65
+ 2025-11-06 18:04:12 - INFO - PyTorch model saved to: /home/philab/Desktop/hydranet/experiments/exp_1762448652_9207_s96_f20_d3_m2_2_2_2/pytorch/model.pt
66
+ 2025-11-06 18:04:12 - INFO - === ONNX Conversion Phase ===
67
+ 2025-11-06 18:04:12 - INFO - === Model Export Diagnostics ===
68
+ 2025-11-06 18:04:12 - INFO - PyTorch version: 1.9.0+cu102
69
+ 2025-11-06 18:04:12 - INFO - Model parameters: 135,363
70
+ 2025-11-06 18:04:12 - INFO - Model memory: 0.52 MB
71
+ 2025-11-06 18:04:12 - INFO - Starting ONNX export with opset version 11
72
+ 2025-11-06 18:04:12 - INFO - Model input shape: torch.Size([1, 8, 96, 96])
73
+ 2025-11-06 18:04:12 - INFO - Model input dtype: torch.float32
74
+ 2025-11-06 18:04:12 - INFO - Forward pass successful. Output shape: torch.Size([1, 3, 96, 96])
75
+ 2025-11-06 18:04:12 - INFO - Output dtype: torch.float32
76
+ 2025-11-06 18:04:12 - INFO - Output value range: [-0.4622, 0.9376]
77
+ 2025-11-06 18:04:12 - INFO - Model successfully exported to /home/philab/Desktop/hydranet/experiments/exp_1762448652_9207_s96_f20_d3_m2_2_2_2/onnx/model.onnx
78
+ 2025-11-06 18:04:12 - INFO - ONNX model size: 0.52 MB
79
+ 2025-11-06 18:04:12 - INFO - Saved dummy input with shape (1, 8, 96, 96) to /home/philab/Desktop/hydranet/experiments/exp_1762448652_9207_s96_f20_d3_m2_2_2_2/onnx/sample_input.npy
80
+ 2025-11-06 18:04:12 - INFO - Input data type: float32
81
+ 2025-11-06 18:04:12 - INFO - Input value range: [-4.0005, 3.9783]
82
+ 2025-11-06 18:04:12 - INFO - === OpenVINO Conversion Phase ===
83
+ 2025-11-06 18:04:12 - INFO - Starting OpenVINO conversion in Docker container...
84
+ 2025-11-06 18:04:16 - INFO - OpenVINO conversion completed in 3.95 seconds
85
+ 2025-11-06 18:04:16 - INFO - OpenVINO model files created:
86
+ 2025-11-06 18:04:16 - INFO - XML file: /home/philab/Desktop/hydranet/experiments/exp_1762448652_9207_s96_f20_d3_m2_2_2_2/openvino/model.xml (0.08 MB)
87
+ 2025-11-06 18:04:16 - INFO - BIN file: /home/philab/Desktop/hydranet/experiments/exp_1762448652_9207_s96_f20_d3_m2_2_2_2/openvino/model.bin (0.26 MB)
88
+ 2025-11-06 18:04:16 - INFO - === Myriad Inference Phase ===
89
+ 2025-11-06 18:04:16 - INFO - Starting Myriad inference in Docker container...
90
+ 2025-11-06 18:04:19 - INFO - Myriad inference completed in 2.77 seconds
91
+ 2025-11-06 18:04:19 - INFO - Actual inference time: 0.148318 seconds
92
+ 2025-11-06 18:04:19 - INFO - ✅ Complete pipeline executed successfully!
93
+ 2025-11-06 18:04:19 - INFO - ✅ Experiment 24 completed successfully
94
+ 2025-11-06 18:04:19 - INFO - Inference time: 0.148318s
95
+ 2025-11-06 18:04:19 - INFO -
96
+ === Experiment 25/2475 ===
97
+ 2025-11-06 18:04:19 - INFO - Experiment ID: exp_1762448659_3550_s96_f20_d3_m1_2_2_2
98
+ 2025-11-06 18:04:19 - INFO - Experiment directory: /home/philab/Desktop/hydranet/experiments/exp_1762448659_3550_s96_f20_d3_m1_2_2_2
exp_1762448652_9207_s96_f20_d3_m2_2_2_2/model_info.json ADDED
@@ -0,0 +1,138 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "config": {
3
+ "input_size": 96,
4
+ "base_filters": 20,
5
+ "depth": 3,
6
+ "channel_multipliers": [
7
+ 2,
8
+ 2,
9
+ 2,
10
+ 2
11
+ ],
12
+ "n_channels": 8,
13
+ "n_classes": 3
14
+ },
15
+ "model_info": {
16
+ "depth": 3,
17
+ "channels_per_level": [
18
+ 40,
19
+ 40,
20
+ 40,
21
+ 40
22
+ ],
23
+ "channel_multipliers": [
24
+ 2,
25
+ 2,
26
+ 2,
27
+ 2
28
+ ],
29
+ "total_parameters": 135363,
30
+ "trainable_parameters": 135363,
31
+ "model_size_mb": 0.5163688659667969
32
+ },
33
+ "architecture_stats": {
34
+ "input_dimension": 73728,
35
+ "output_dimension": 27648,
36
+ "parameter_distribution": {
37
+ "encoder_params": 45600,
38
+ "decoder_params": 74280,
39
+ "bottleneck_params": 15080,
40
+ "other_params": 123,
41
+ "total_params": 135083,
42
+ "encoder_percentage": 33.75702345965073,
43
+ "decoder_percentage": 54.988414530325805,
44
+ "bottleneck_percentage": 11.163506880954673,
45
+ "other_percentage": 0.09105512906879475
46
+ },
47
+ "component_breakdown": {
48
+ "Encoder - encoders.0.channel_proj": 360,
49
+ "Encoder - encoders.0.convnext_block.dwconv": 2000,
50
+ "Encoder - encoders.0.convnext_block.norm": 80,
51
+ "Encoder - encoders.0.convnext_block.pwconv1": 6560,
52
+ "Encoder - encoders.0.convnext_block.pwconv2": 6440,
53
+ "Encoder - encoders.1.convnext_block.dwconv": 2000,
54
+ "Encoder - encoders.1.convnext_block.norm": 80,
55
+ "Encoder - encoders.1.convnext_block.pwconv1": 6560,
56
+ "Encoder - encoders.1.convnext_block.pwconv2": 6440,
57
+ "Encoder - encoders.2.convnext_block.dwconv": 2000,
58
+ "Encoder - encoders.2.convnext_block.norm": 80,
59
+ "Encoder - encoders.2.convnext_block.pwconv1": 6560,
60
+ "Encoder - encoders.2.convnext_block.pwconv2": 6440,
61
+ "Bottleneck - bottleneck.convnext_block.dwconv": 2000,
62
+ "Bottleneck - bottleneck.convnext_block.norm": 80,
63
+ "Bottleneck - bottleneck.convnext_block.pwconv1": 6560,
64
+ "Bottleneck - bottleneck.convnext_block.pwconv2": 6440,
65
+ "Decoder - upsamplers.0": 6440,
66
+ "Decoder - upsamplers.1": 6440,
67
+ "Decoder - upsamplers.2": 6440,
68
+ "Decoder - decoders.0.channel_proj": 3240,
69
+ "Decoder - decoders.0.convnext_block.dwconv": 2000,
70
+ "Decoder - decoders.0.convnext_block.norm": 80,
71
+ "Decoder - decoders.0.convnext_block.pwconv1": 6560,
72
+ "Decoder - decoders.0.convnext_block.pwconv2": 6440,
73
+ "Decoder - decoders.1.channel_proj": 3240,
74
+ "Decoder - decoders.1.convnext_block.dwconv": 2000,
75
+ "Decoder - decoders.1.convnext_block.norm": 80,
76
+ "Decoder - decoders.1.convnext_block.pwconv1": 6560,
77
+ "Decoder - decoders.1.convnext_block.pwconv2": 6440,
78
+ "Decoder - decoders.2.channel_proj": 3240,
79
+ "Decoder - decoders.2.convnext_block.dwconv": 2000,
80
+ "Decoder - decoders.2.convnext_block.norm": 80,
81
+ "Decoder - decoders.2.convnext_block.pwconv1": 6560,
82
+ "Decoder - decoders.2.convnext_block.pwconv2": 6440,
83
+ "Other - final_conv": 123
84
+ },
85
+ "latent_dimensions": {
86
+ "Level_0": {
87
+ "channels": 40,
88
+ "height": 96,
89
+ "width": 96,
90
+ "total_elements": 368640
91
+ },
92
+ "Level_1": {
93
+ "channels": 40,
94
+ "height": 48,
95
+ "width": 48,
96
+ "total_elements": 92160
97
+ },
98
+ "Level_2": {
99
+ "channels": 40,
100
+ "height": 24,
101
+ "width": 24,
102
+ "total_elements": 23040
103
+ },
104
+ "Level_3": {
105
+ "channels": 40,
106
+ "height": 12,
107
+ "width": 12,
108
+ "total_elements": 5760
109
+ }
110
+ },
111
+ "skip_dimensions": {
112
+ "Skip_0": {
113
+ "channels": 40,
114
+ "height": 96,
115
+ "width": 96,
116
+ "total_elements": 368640
117
+ },
118
+ "Skip_1": {
119
+ "channels": 40,
120
+ "height": 48,
121
+ "width": 48,
122
+ "total_elements": 92160
123
+ },
124
+ "Skip_2": {
125
+ "channels": 40,
126
+ "height": 24,
127
+ "width": 24,
128
+ "total_elements": 23040
129
+ }
130
+ },
131
+ "memory_analysis": {
132
+ "peak_memory_inference_mb": 2.14892578125,
133
+ "peak_memory_training_mb": 4.2978515625,
134
+ "peak_elements": 563328
135
+ }
136
+ },
137
+ "timestamp": "2025-11-06T18:04:12.573784"
138
+ }
exp_1762448652_9207_s96_f20_d3_m2_2_2_2/pipeline_results.json ADDED
@@ -0,0 +1,138 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "experiment_id": "exp_1762448652_6643_s96_f20_d3_m2_2_2_2",
3
+ "config": {
4
+ "input_size": 96,
5
+ "base_filters": 20,
6
+ "depth": 3,
7
+ "channel_multipliers": [
8
+ 2,
9
+ 2,
10
+ 2,
11
+ 2
12
+ ],
13
+ "n_channels": 8,
14
+ "n_classes": 3
15
+ },
16
+ "start_time": "2025-11-06T18:04:12.566367",
17
+ "success": true,
18
+ "timings": {
19
+ "onnx_conversion": 0.26931651402264833,
20
+ "openvino_conversion": 3.9536154859233648,
21
+ "inference_total": 2.7693776809610426,
22
+ "inference_actual": 0.148318,
23
+ "total": 7.005332896020263
24
+ },
25
+ "errors": [],
26
+ "architecture_stats": {
27
+ "input_dimension": 73728,
28
+ "output_dimension": 27648,
29
+ "parameter_distribution": {
30
+ "encoder_params": 45600,
31
+ "decoder_params": 74280,
32
+ "bottleneck_params": 15080,
33
+ "other_params": 123,
34
+ "total_params": 135083,
35
+ "encoder_percentage": 33.75702345965073,
36
+ "decoder_percentage": 54.988414530325805,
37
+ "bottleneck_percentage": 11.163506880954673,
38
+ "other_percentage": 0.09105512906879475
39
+ },
40
+ "component_breakdown": {
41
+ "Encoder - encoders.0.channel_proj": 360,
42
+ "Encoder - encoders.0.convnext_block.dwconv": 2000,
43
+ "Encoder - encoders.0.convnext_block.norm": 80,
44
+ "Encoder - encoders.0.convnext_block.pwconv1": 6560,
45
+ "Encoder - encoders.0.convnext_block.pwconv2": 6440,
46
+ "Encoder - encoders.1.convnext_block.dwconv": 2000,
47
+ "Encoder - encoders.1.convnext_block.norm": 80,
48
+ "Encoder - encoders.1.convnext_block.pwconv1": 6560,
49
+ "Encoder - encoders.1.convnext_block.pwconv2": 6440,
50
+ "Encoder - encoders.2.convnext_block.dwconv": 2000,
51
+ "Encoder - encoders.2.convnext_block.norm": 80,
52
+ "Encoder - encoders.2.convnext_block.pwconv1": 6560,
53
+ "Encoder - encoders.2.convnext_block.pwconv2": 6440,
54
+ "Bottleneck - bottleneck.convnext_block.dwconv": 2000,
55
+ "Bottleneck - bottleneck.convnext_block.norm": 80,
56
+ "Bottleneck - bottleneck.convnext_block.pwconv1": 6560,
57
+ "Bottleneck - bottleneck.convnext_block.pwconv2": 6440,
58
+ "Decoder - upsamplers.0": 6440,
59
+ "Decoder - upsamplers.1": 6440,
60
+ "Decoder - upsamplers.2": 6440,
61
+ "Decoder - decoders.0.channel_proj": 3240,
62
+ "Decoder - decoders.0.convnext_block.dwconv": 2000,
63
+ "Decoder - decoders.0.convnext_block.norm": 80,
64
+ "Decoder - decoders.0.convnext_block.pwconv1": 6560,
65
+ "Decoder - decoders.0.convnext_block.pwconv2": 6440,
66
+ "Decoder - decoders.1.channel_proj": 3240,
67
+ "Decoder - decoders.1.convnext_block.dwconv": 2000,
68
+ "Decoder - decoders.1.convnext_block.norm": 80,
69
+ "Decoder - decoders.1.convnext_block.pwconv1": 6560,
70
+ "Decoder - decoders.1.convnext_block.pwconv2": 6440,
71
+ "Decoder - decoders.2.channel_proj": 3240,
72
+ "Decoder - decoders.2.convnext_block.dwconv": 2000,
73
+ "Decoder - decoders.2.convnext_block.norm": 80,
74
+ "Decoder - decoders.2.convnext_block.pwconv1": 6560,
75
+ "Decoder - decoders.2.convnext_block.pwconv2": 6440,
76
+ "Other - final_conv": 123
77
+ },
78
+ "latent_dimensions": {
79
+ "Level_0": {
80
+ "channels": 40,
81
+ "height": 96,
82
+ "width": 96,
83
+ "total_elements": 368640
84
+ },
85
+ "Level_1": {
86
+ "channels": 40,
87
+ "height": 48,
88
+ "width": 48,
89
+ "total_elements": 92160
90
+ },
91
+ "Level_2": {
92
+ "channels": 40,
93
+ "height": 24,
94
+ "width": 24,
95
+ "total_elements": 23040
96
+ },
97
+ "Level_3": {
98
+ "channels": 40,
99
+ "height": 12,
100
+ "width": 12,
101
+ "total_elements": 5760
102
+ }
103
+ },
104
+ "skip_dimensions": {
105
+ "Skip_0": {
106
+ "channels": 40,
107
+ "height": 96,
108
+ "width": 96,
109
+ "total_elements": 368640
110
+ },
111
+ "Skip_1": {
112
+ "channels": 40,
113
+ "height": 48,
114
+ "width": 48,
115
+ "total_elements": 92160
116
+ },
117
+ "Skip_2": {
118
+ "channels": 40,
119
+ "height": 24,
120
+ "width": 24,
121
+ "total_elements": 23040
122
+ }
123
+ },
124
+ "memory_analysis": {
125
+ "peak_memory_inference_mb": 2.14892578125,
126
+ "peak_memory_training_mb": 4.2978515625,
127
+ "peak_elements": 563328
128
+ }
129
+ },
130
+ "inference_results": {
131
+ "success": true,
132
+ "total_time": 2.7693776809610426,
133
+ "inference_time": 0.148318,
134
+ "stdout": "[setupvars.sh] OpenVINO environment initialized\nStarting Myriad inference...\nInput: /home/mount/experiments/exp_1762448652_9207_s96_f20_d3_m2_2_2_2/onnx/sample_input.npy\nModel XML: /home/mount/experiments/exp_1762448652_9207_s96_f20_d3_m2_2_2_2/openvino/model.xml\nModel BIN: /home/mount/experiments/exp_1762448652_9207_s96_f20_d3_m2_2_2_2/openvino/model.bin\n[OK] OpenVINO inference engine imported successfully\nDevice: MYRIAD\nModel XML: /home/mount/experiments/exp_1762448652_9207_s96_f20_d3_m2_2_2_2/openvino/model.xml\nModel BIN: /home/mount/experiments/exp_1762448652_9207_s96_f20_d3_m2_2_2_2/openvino/model.bin\nInput file: /home/mount/experiments/exp_1762448652_9207_s96_f20_d3_m2_2_2_2/onnx/sample_input.npy\nInitializing OpenVINO Runtime Core...\nAvailable devices:\n[E:] [BSL] found 0 ioexpander device\n ['CPU', 'GNA', 'MYRIAD']\nLoading network...\nInput blob: input\nInput shape: [1, 8, 96, 96]\nOutput blob: Conv_104\nOutput shape: [1, 3, 96, 96]\nLoading network to MYRIAD...\nLoading input data...\nInput data shape: (1, 8, 96, 96)\nInput data type: float32\nRunning inference...\n[OK] Inference completed!\nInference time: 0.148318 seconds\nOutput shape: (1, 3, 96, 96)\nOutput dtype: float32\nOutput range: [-0.462646, 0.937988]\nMyriad inference completed!\n",
135
+ "stderr": ""
136
+ },
137
+ "end_time": "2025-11-06T18:04:19.571690"
138
+ }
exp_1762448652_9207_s96_f20_d3_m2_2_2_2/run_inference.sh ADDED
@@ -0,0 +1,19 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #!/bin/bash
2
+ # Source OpenVINO environment
3
+ source /opt/intel/openvino/bin/setupvars.sh
4
+
5
+ PYPATH=/home/mount/scripts/simple_inference.py
6
+
7
+ echo "Starting Myriad inference..."
8
+ echo "Input: /home/mount/experiments/exp_1762448652_9207_s96_f20_d3_m2_2_2_2/onnx/sample_input.npy"
9
+ echo "Model XML: /home/mount/experiments/exp_1762448652_9207_s96_f20_d3_m2_2_2_2/openvino/model.xml"
10
+ echo "Model BIN: /home/mount/experiments/exp_1762448652_9207_s96_f20_d3_m2_2_2_2/openvino/model.bin"
11
+
12
+ python3 $PYPATH \
13
+ --input_filepath /home/mount/experiments/exp_1762448652_9207_s96_f20_d3_m2_2_2_2/onnx/sample_input.npy \
14
+ --device_name MYRIAD \
15
+ --model_bin /home/mount/experiments/exp_1762448652_9207_s96_f20_d3_m2_2_2_2/openvino/model.bin \
16
+ --model_xml /home/mount/experiments/exp_1762448652_9207_s96_f20_d3_m2_2_2_2/openvino/model.xml \
17
+ --output /home/mount/experiments/exp_1762448652_9207_s96_f20_d3_m2_2_2_2/inference/
18
+
19
+ echo "Myriad inference completed!"
exp_1762448723_5984_s96_f22_d3_m1_1_1_1/convert_openvino.sh ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #!/bin/bash
2
+ # Source OpenVINO environment
3
+ source /opt/intel/openvino/bin/setupvars.sh
4
+
5
+ # OpenVINO Model Optimizer script for ONNX model
6
+ PYPATH=/opt/intel/openvino_2020.3.194/deployment_tools/model_optimizer/mo.py
7
+ ONNX_MODEL=/home/mount/experiments/exp_1762448723_5984_s96_f22_d3_m1_1_1_1/onnx/model.onnx
8
+ OUTPUT_DIR=/home/mount/experiments/exp_1762448723_5984_s96_f22_d3_m1_1_1_1/openvino
9
+
10
+ echo "Converting ONNX model to OpenVINO IR format..."
11
+ echo "Input model: $ONNX_MODEL"
12
+ echo "Output directory: $OUTPUT_DIR"
13
+
14
+ # Create output directory if it doesn't exist
15
+ mkdir -p "$OUTPUT_DIR"
16
+
17
+ python3 $PYPATH \
18
+ --input_model "$ONNX_MODEL" \
19
+ --data_type FP16 \
20
+ --input_shape "[1,8,96,96]" \
21
+ --mean_values "[0,0,0,0,0,0,0,0]" \
22
+ --scale_values "[1,1,1,1,1,1,1,1]" \
23
+ --progress \
24
+ --stream_output \
25
+ --output_dir "$OUTPUT_DIR" \
26
+ --model_name model
27
+
28
+ echo "OpenVINO conversion completed!"
29
+ echo "Generated files in $OUTPUT_DIR:"
30
+ ls -la "$OUTPUT_DIR/"
exp_1762448723_5984_s96_f22_d3_m1_1_1_1/experiment.log ADDED
@@ -0,0 +1,98 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2025-11-06 18:05:23 - INFO - === Model Creation Phase ===
2
+ 2025-11-06 18:05:23 - INFO - Creating UNet model with config: {'input_size': 96, 'base_filters': 22, 'depth': 3, 'channel_multipliers': [1, 1, 1, 1], 'n_channels': 8, 'n_classes': 3}
3
+ 2025-11-06 18:05:23 - INFO - Model created successfully:
4
+ 2025-11-06 18:05:23 - INFO - Depth: 3
5
+ 2025-11-06 18:05:23 - INFO - Channels per level: [22, 22, 22, 22]
6
+ 2025-11-06 18:05:23 - INFO - Total parameters: 45,147
7
+ 2025-11-06 18:05:23 - INFO - Parameter memory: 0.17 MB
8
+ 2025-11-06 18:05:23 - INFO - === Detailed Architecture Analysis ===
9
+ 2025-11-06 18:05:23 - INFO - Input vector dimension: 73,728 (8 × 96 × 96)
10
+ 2025-11-06 18:05:23 - INFO - Component-wise parameter breakdown:
11
+ 2025-11-06 18:05:23 - INFO - Encoder - encoders.0.channel_proj: 198 parameters
12
+ 2025-11-06 18:05:23 - INFO - Encoder - encoders.0.convnext_block.dwconv: 1,100 parameters
13
+ 2025-11-06 18:05:23 - INFO - Encoder - encoders.0.convnext_block.norm: 44 parameters
14
+ 2025-11-06 18:05:23 - INFO - Encoder - encoders.0.convnext_block.pwconv1: 2,024 parameters
15
+ 2025-11-06 18:05:23 - INFO - Encoder - encoders.0.convnext_block.pwconv2: 1,958 parameters
16
+ 2025-11-06 18:05:23 - INFO - Encoder - encoders.1.convnext_block.dwconv: 1,100 parameters
17
+ 2025-11-06 18:05:23 - INFO - Encoder - encoders.1.convnext_block.norm: 44 parameters
18
+ 2025-11-06 18:05:23 - INFO - Encoder - encoders.1.convnext_block.pwconv1: 2,024 parameters
19
+ 2025-11-06 18:05:23 - INFO - Encoder - encoders.1.convnext_block.pwconv2: 1,958 parameters
20
+ 2025-11-06 18:05:23 - INFO - Encoder - encoders.2.convnext_block.dwconv: 1,100 parameters
21
+ 2025-11-06 18:05:23 - INFO - Encoder - encoders.2.convnext_block.norm: 44 parameters
22
+ 2025-11-06 18:05:23 - INFO - Encoder - encoders.2.convnext_block.pwconv1: 2,024 parameters
23
+ 2025-11-06 18:05:23 - INFO - Encoder - encoders.2.convnext_block.pwconv2: 1,958 parameters
24
+ 2025-11-06 18:05:23 - INFO - Bottleneck - bottleneck.convnext_block.dwconv: 1,100 parameters
25
+ 2025-11-06 18:05:23 - INFO - Bottleneck - bottleneck.convnext_block.norm: 44 parameters
26
+ 2025-11-06 18:05:23 - INFO - Bottleneck - bottleneck.convnext_block.pwconv1: 2,024 parameters
27
+ 2025-11-06 18:05:23 - INFO - Bottleneck - bottleneck.convnext_block.pwconv2: 1,958 parameters
28
+ 2025-11-06 18:05:23 - INFO - Decoder - upsamplers.0: 1,958 parameters
29
+ 2025-11-06 18:05:23 - INFO - Decoder - upsamplers.1: 1,958 parameters
30
+ 2025-11-06 18:05:23 - INFO - Decoder - upsamplers.2: 1,958 parameters
31
+ 2025-11-06 18:05:23 - INFO - Decoder - decoders.0.channel_proj: 990 parameters
32
+ 2025-11-06 18:05:23 - INFO - Decoder - decoders.0.convnext_block.dwconv: 1,100 parameters
33
+ 2025-11-06 18:05:23 - INFO - Decoder - decoders.0.convnext_block.norm: 44 parameters
34
+ 2025-11-06 18:05:23 - INFO - Decoder - decoders.0.convnext_block.pwconv1: 2,024 parameters
35
+ 2025-11-06 18:05:23 - INFO - Decoder - decoders.0.convnext_block.pwconv2: 1,958 parameters
36
+ 2025-11-06 18:05:23 - INFO - Decoder - decoders.1.channel_proj: 990 parameters
37
+ 2025-11-06 18:05:23 - INFO - Decoder - decoders.1.convnext_block.dwconv: 1,100 parameters
38
+ 2025-11-06 18:05:23 - INFO - Decoder - decoders.1.convnext_block.norm: 44 parameters
39
+ 2025-11-06 18:05:23 - INFO - Decoder - decoders.1.convnext_block.pwconv1: 2,024 parameters
40
+ 2025-11-06 18:05:23 - INFO - Decoder - decoders.1.convnext_block.pwconv2: 1,958 parameters
41
+ 2025-11-06 18:05:23 - INFO - Decoder - decoders.2.channel_proj: 990 parameters
42
+ 2025-11-06 18:05:23 - INFO - Decoder - decoders.2.convnext_block.dwconv: 1,100 parameters
43
+ 2025-11-06 18:05:23 - INFO - Decoder - decoders.2.convnext_block.norm: 44 parameters
44
+ 2025-11-06 18:05:23 - INFO - Decoder - decoders.2.convnext_block.pwconv1: 2,024 parameters
45
+ 2025-11-06 18:05:23 - INFO - Decoder - decoders.2.convnext_block.pwconv2: 1,958 parameters
46
+ 2025-11-06 18:05:23 - INFO - Other - final_conv: 69 parameters
47
+ 2025-11-06 18:05:23 - INFO - Parameter distribution summary:
48
+ 2025-11-06 18:05:23 - INFO - Encoder parameters: 15,576 (34.6%)
49
+ 2025-11-06 18:05:23 - INFO - Decoder parameters: 24,222 (53.8%)
50
+ 2025-11-06 18:05:23 - INFO - Bottleneck parameters: 5,126 (11.4%)
51
+ 2025-11-06 18:05:23 - INFO - Other parameters: 69 (0.2%)
52
+ 2025-11-06 18:05:23 - INFO - Latent space dimensions (feature maps at each level):
53
+ 2025-11-06 18:05:23 - INFO - Level 0: 22 × 96 × 96 = 202,752 elements
54
+ 2025-11-06 18:05:23 - INFO - Level 1: 22 × 48 × 48 = 50,688 elements
55
+ 2025-11-06 18:05:23 - INFO - Level 2: 22 × 24 × 24 = 12,672 elements
56
+ 2025-11-06 18:05:23 - INFO - Level 3: 22 × 12 × 12 = 3,168 elements
57
+ 2025-11-06 18:05:23 - INFO - Skip connection dimensions:
58
+ 2025-11-06 18:05:23 - INFO - Skip 0: 22 × 96 × 96 = 202,752 elements
59
+ 2025-11-06 18:05:23 - INFO - Skip 1: 22 × 48 × 48 = 50,688 elements
60
+ 2025-11-06 18:05:23 - INFO - Skip 2: 22 × 24 × 24 = 12,672 elements
61
+ 2025-11-06 18:05:23 - INFO - Memory analysis:
62
+ 2025-11-06 18:05:23 - INFO - Peak feature map memory (inference): 1.31 MB
63
+ 2025-11-06 18:05:23 - INFO - Peak feature map memory (training): 2.62 MB (with gradients)
64
+ 2025-11-06 18:05:23 - INFO - Output vector dimension: 27,648 (3 × 96 × 96)
65
+ 2025-11-06 18:05:23 - INFO - PyTorch model saved to: /home/philab/Desktop/hydranet/experiments/exp_1762448723_5984_s96_f22_d3_m1_1_1_1/pytorch/model.pt
66
+ 2025-11-06 18:05:23 - INFO - === ONNX Conversion Phase ===
67
+ 2025-11-06 18:05:23 - INFO - === Model Export Diagnostics ===
68
+ 2025-11-06 18:05:23 - INFO - PyTorch version: 1.9.0+cu102
69
+ 2025-11-06 18:05:23 - INFO - Model parameters: 45,147
70
+ 2025-11-06 18:05:23 - INFO - Model memory: 0.17 MB
71
+ 2025-11-06 18:05:23 - INFO - Starting ONNX export with opset version 11
72
+ 2025-11-06 18:05:23 - INFO - Model input shape: torch.Size([1, 8, 96, 96])
73
+ 2025-11-06 18:05:23 - INFO - Model input dtype: torch.float32
74
+ 2025-11-06 18:05:23 - INFO - Forward pass successful. Output shape: torch.Size([1, 3, 96, 96])
75
+ 2025-11-06 18:05:23 - INFO - Output dtype: torch.float32
76
+ 2025-11-06 18:05:23 - INFO - Output value range: [-0.6694, 0.8579]
77
+ 2025-11-06 18:05:24 - INFO - Model successfully exported to /home/philab/Desktop/hydranet/experiments/exp_1762448723_5984_s96_f22_d3_m1_1_1_1/onnx/model.onnx
78
+ 2025-11-06 18:05:24 - INFO - ONNX model size: 0.18 MB
79
+ 2025-11-06 18:05:24 - INFO - Saved dummy input with shape (1, 8, 96, 96) to /home/philab/Desktop/hydranet/experiments/exp_1762448723_5984_s96_f22_d3_m1_1_1_1/onnx/sample_input.npy
80
+ 2025-11-06 18:05:24 - INFO - Input data type: float32
81
+ 2025-11-06 18:05:24 - INFO - Input value range: [-4.4875, 4.1861]
82
+ 2025-11-06 18:05:24 - INFO - === OpenVINO Conversion Phase ===
83
+ 2025-11-06 18:05:24 - INFO - Starting OpenVINO conversion in Docker container...
84
+ 2025-11-06 18:05:27 - INFO - OpenVINO conversion completed in 3.91 seconds
85
+ 2025-11-06 18:05:27 - INFO - OpenVINO model files created:
86
+ 2025-11-06 18:05:27 - INFO - XML file: /home/philab/Desktop/hydranet/experiments/exp_1762448723_5984_s96_f22_d3_m1_1_1_1/openvino/model.xml (0.08 MB)
87
+ 2025-11-06 18:05:27 - INFO - BIN file: /home/philab/Desktop/hydranet/experiments/exp_1762448723_5984_s96_f22_d3_m1_1_1_1/openvino/model.bin (0.09 MB)
88
+ 2025-11-06 18:05:27 - INFO - === Myriad Inference Phase ===
89
+ 2025-11-06 18:05:27 - INFO - Starting Myriad inference in Docker container...
90
+ 2025-11-06 18:05:30 - INFO - Myriad inference completed in 2.77 seconds
91
+ 2025-11-06 18:05:30 - INFO - Actual inference time: 0.175996 seconds
92
+ 2025-11-06 18:05:30 - INFO - ✅ Complete pipeline executed successfully!
93
+ 2025-11-06 18:05:30 - INFO - ✅ Experiment 34 completed successfully
94
+ 2025-11-06 18:05:30 - INFO - Inference time: 0.175996s
95
+ 2025-11-06 18:05:30 - INFO -
96
+ === Experiment 35/2475 ===
97
+ 2025-11-06 18:05:30 - INFO - Experiment ID: exp_1762448730_3552_s96_f22_d3_m2_2_2_2
98
+ 2025-11-06 18:05:30 - INFO - Experiment directory: /home/philab/Desktop/hydranet/experiments/exp_1762448730_3552_s96_f22_d3_m2_2_2_2
exp_1762448723_5984_s96_f22_d3_m1_1_1_1/model_info.json ADDED
@@ -0,0 +1,138 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "config": {
3
+ "input_size": 96,
4
+ "base_filters": 22,
5
+ "depth": 3,
6
+ "channel_multipliers": [
7
+ 1,
8
+ 1,
9
+ 1,
10
+ 1
11
+ ],
12
+ "n_channels": 8,
13
+ "n_classes": 3
14
+ },
15
+ "model_info": {
16
+ "depth": 3,
17
+ "channels_per_level": [
18
+ 22,
19
+ 22,
20
+ 22,
21
+ 22
22
+ ],
23
+ "channel_multipliers": [
24
+ 1,
25
+ 1,
26
+ 1,
27
+ 1
28
+ ],
29
+ "total_parameters": 45147,
30
+ "trainable_parameters": 45147,
31
+ "model_size_mb": 0.17222213745117188
32
+ },
33
+ "architecture_stats": {
34
+ "input_dimension": 73728,
35
+ "output_dimension": 27648,
36
+ "parameter_distribution": {
37
+ "encoder_params": 15576,
38
+ "decoder_params": 24222,
39
+ "bottleneck_params": 5126,
40
+ "other_params": 69,
41
+ "total_params": 44993,
42
+ "encoder_percentage": 34.61871846731714,
43
+ "decoder_percentage": 53.83504100637877,
44
+ "bottleneck_percentage": 11.39288333740804,
45
+ "other_percentage": 0.1533571888960505
46
+ },
47
+ "component_breakdown": {
48
+ "Encoder - encoders.0.channel_proj": 198,
49
+ "Encoder - encoders.0.convnext_block.dwconv": 1100,
50
+ "Encoder - encoders.0.convnext_block.norm": 44,
51
+ "Encoder - encoders.0.convnext_block.pwconv1": 2024,
52
+ "Encoder - encoders.0.convnext_block.pwconv2": 1958,
53
+ "Encoder - encoders.1.convnext_block.dwconv": 1100,
54
+ "Encoder - encoders.1.convnext_block.norm": 44,
55
+ "Encoder - encoders.1.convnext_block.pwconv1": 2024,
56
+ "Encoder - encoders.1.convnext_block.pwconv2": 1958,
57
+ "Encoder - encoders.2.convnext_block.dwconv": 1100,
58
+ "Encoder - encoders.2.convnext_block.norm": 44,
59
+ "Encoder - encoders.2.convnext_block.pwconv1": 2024,
60
+ "Encoder - encoders.2.convnext_block.pwconv2": 1958,
61
+ "Bottleneck - bottleneck.convnext_block.dwconv": 1100,
62
+ "Bottleneck - bottleneck.convnext_block.norm": 44,
63
+ "Bottleneck - bottleneck.convnext_block.pwconv1": 2024,
64
+ "Bottleneck - bottleneck.convnext_block.pwconv2": 1958,
65
+ "Decoder - upsamplers.0": 1958,
66
+ "Decoder - upsamplers.1": 1958,
67
+ "Decoder - upsamplers.2": 1958,
68
+ "Decoder - decoders.0.channel_proj": 990,
69
+ "Decoder - decoders.0.convnext_block.dwconv": 1100,
70
+ "Decoder - decoders.0.convnext_block.norm": 44,
71
+ "Decoder - decoders.0.convnext_block.pwconv1": 2024,
72
+ "Decoder - decoders.0.convnext_block.pwconv2": 1958,
73
+ "Decoder - decoders.1.channel_proj": 990,
74
+ "Decoder - decoders.1.convnext_block.dwconv": 1100,
75
+ "Decoder - decoders.1.convnext_block.norm": 44,
76
+ "Decoder - decoders.1.convnext_block.pwconv1": 2024,
77
+ "Decoder - decoders.1.convnext_block.pwconv2": 1958,
78
+ "Decoder - decoders.2.channel_proj": 990,
79
+ "Decoder - decoders.2.convnext_block.dwconv": 1100,
80
+ "Decoder - decoders.2.convnext_block.norm": 44,
81
+ "Decoder - decoders.2.convnext_block.pwconv1": 2024,
82
+ "Decoder - decoders.2.convnext_block.pwconv2": 1958,
83
+ "Other - final_conv": 69
84
+ },
85
+ "latent_dimensions": {
86
+ "Level_0": {
87
+ "channels": 22,
88
+ "height": 96,
89
+ "width": 96,
90
+ "total_elements": 202752
91
+ },
92
+ "Level_1": {
93
+ "channels": 22,
94
+ "height": 48,
95
+ "width": 48,
96
+ "total_elements": 50688
97
+ },
98
+ "Level_2": {
99
+ "channels": 22,
100
+ "height": 24,
101
+ "width": 24,
102
+ "total_elements": 12672
103
+ },
104
+ "Level_3": {
105
+ "channels": 22,
106
+ "height": 12,
107
+ "width": 12,
108
+ "total_elements": 3168
109
+ }
110
+ },
111
+ "skip_dimensions": {
112
+ "Skip_0": {
113
+ "channels": 22,
114
+ "height": 96,
115
+ "width": 96,
116
+ "total_elements": 202752
117
+ },
118
+ "Skip_1": {
119
+ "channels": 22,
120
+ "height": 48,
121
+ "width": 48,
122
+ "total_elements": 50688
123
+ },
124
+ "Skip_2": {
125
+ "channels": 22,
126
+ "height": 24,
127
+ "width": 24,
128
+ "total_elements": 12672
129
+ }
130
+ },
131
+ "memory_analysis": {
132
+ "peak_memory_inference_mb": 1.3084716796875,
133
+ "peak_memory_training_mb": 2.616943359375,
134
+ "peak_elements": 343008
135
+ }
136
+ },
137
+ "timestamp": "2025-11-06T18:05:23.754547"
138
+ }
exp_1762448723_5984_s96_f22_d3_m1_1_1_1/pipeline_results.json ADDED
@@ -0,0 +1,138 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "experiment_id": "exp_1762448723_5692_s96_f22_d3_m1_1_1_1",
3
+ "config": {
4
+ "input_size": 96,
5
+ "base_filters": 22,
6
+ "depth": 3,
7
+ "channel_multipliers": [
8
+ 1,
9
+ 1,
10
+ 1,
11
+ 1
12
+ ],
13
+ "n_channels": 8,
14
+ "n_classes": 3
15
+ },
16
+ "start_time": "2025-11-06T18:05:23.748640",
17
+ "success": true,
18
+ "timings": {
19
+ "onnx_conversion": 0.25980374589562416,
20
+ "openvino_conversion": 3.912094983039424,
21
+ "inference_total": 2.765632274094969,
22
+ "inference_actual": 0.175996,
23
+ "total": 6.948178299935535
24
+ },
25
+ "errors": [],
26
+ "architecture_stats": {
27
+ "input_dimension": 73728,
28
+ "output_dimension": 27648,
29
+ "parameter_distribution": {
30
+ "encoder_params": 15576,
31
+ "decoder_params": 24222,
32
+ "bottleneck_params": 5126,
33
+ "other_params": 69,
34
+ "total_params": 44993,
35
+ "encoder_percentage": 34.61871846731714,
36
+ "decoder_percentage": 53.83504100637877,
37
+ "bottleneck_percentage": 11.39288333740804,
38
+ "other_percentage": 0.1533571888960505
39
+ },
40
+ "component_breakdown": {
41
+ "Encoder - encoders.0.channel_proj": 198,
42
+ "Encoder - encoders.0.convnext_block.dwconv": 1100,
43
+ "Encoder - encoders.0.convnext_block.norm": 44,
44
+ "Encoder - encoders.0.convnext_block.pwconv1": 2024,
45
+ "Encoder - encoders.0.convnext_block.pwconv2": 1958,
46
+ "Encoder - encoders.1.convnext_block.dwconv": 1100,
47
+ "Encoder - encoders.1.convnext_block.norm": 44,
48
+ "Encoder - encoders.1.convnext_block.pwconv1": 2024,
49
+ "Encoder - encoders.1.convnext_block.pwconv2": 1958,
50
+ "Encoder - encoders.2.convnext_block.dwconv": 1100,
51
+ "Encoder - encoders.2.convnext_block.norm": 44,
52
+ "Encoder - encoders.2.convnext_block.pwconv1": 2024,
53
+ "Encoder - encoders.2.convnext_block.pwconv2": 1958,
54
+ "Bottleneck - bottleneck.convnext_block.dwconv": 1100,
55
+ "Bottleneck - bottleneck.convnext_block.norm": 44,
56
+ "Bottleneck - bottleneck.convnext_block.pwconv1": 2024,
57
+ "Bottleneck - bottleneck.convnext_block.pwconv2": 1958,
58
+ "Decoder - upsamplers.0": 1958,
59
+ "Decoder - upsamplers.1": 1958,
60
+ "Decoder - upsamplers.2": 1958,
61
+ "Decoder - decoders.0.channel_proj": 990,
62
+ "Decoder - decoders.0.convnext_block.dwconv": 1100,
63
+ "Decoder - decoders.0.convnext_block.norm": 44,
64
+ "Decoder - decoders.0.convnext_block.pwconv1": 2024,
65
+ "Decoder - decoders.0.convnext_block.pwconv2": 1958,
66
+ "Decoder - decoders.1.channel_proj": 990,
67
+ "Decoder - decoders.1.convnext_block.dwconv": 1100,
68
+ "Decoder - decoders.1.convnext_block.norm": 44,
69
+ "Decoder - decoders.1.convnext_block.pwconv1": 2024,
70
+ "Decoder - decoders.1.convnext_block.pwconv2": 1958,
71
+ "Decoder - decoders.2.channel_proj": 990,
72
+ "Decoder - decoders.2.convnext_block.dwconv": 1100,
73
+ "Decoder - decoders.2.convnext_block.norm": 44,
74
+ "Decoder - decoders.2.convnext_block.pwconv1": 2024,
75
+ "Decoder - decoders.2.convnext_block.pwconv2": 1958,
76
+ "Other - final_conv": 69
77
+ },
78
+ "latent_dimensions": {
79
+ "Level_0": {
80
+ "channels": 22,
81
+ "height": 96,
82
+ "width": 96,
83
+ "total_elements": 202752
84
+ },
85
+ "Level_1": {
86
+ "channels": 22,
87
+ "height": 48,
88
+ "width": 48,
89
+ "total_elements": 50688
90
+ },
91
+ "Level_2": {
92
+ "channels": 22,
93
+ "height": 24,
94
+ "width": 24,
95
+ "total_elements": 12672
96
+ },
97
+ "Level_3": {
98
+ "channels": 22,
99
+ "height": 12,
100
+ "width": 12,
101
+ "total_elements": 3168
102
+ }
103
+ },
104
+ "skip_dimensions": {
105
+ "Skip_0": {
106
+ "channels": 22,
107
+ "height": 96,
108
+ "width": 96,
109
+ "total_elements": 202752
110
+ },
111
+ "Skip_1": {
112
+ "channels": 22,
113
+ "height": 48,
114
+ "width": 48,
115
+ "total_elements": 50688
116
+ },
117
+ "Skip_2": {
118
+ "channels": 22,
119
+ "height": 24,
120
+ "width": 24,
121
+ "total_elements": 12672
122
+ }
123
+ },
124
+ "memory_analysis": {
125
+ "peak_memory_inference_mb": 1.3084716796875,
126
+ "peak_memory_training_mb": 2.616943359375,
127
+ "peak_elements": 343008
128
+ }
129
+ },
130
+ "inference_results": {
131
+ "success": true,
132
+ "total_time": 2.765632274094969,
133
+ "inference_time": 0.175996,
134
+ "stdout": "[setupvars.sh] OpenVINO environment initialized\nStarting Myriad inference...\nInput: /home/mount/experiments/exp_1762448723_5984_s96_f22_d3_m1_1_1_1/onnx/sample_input.npy\nModel XML: /home/mount/experiments/exp_1762448723_5984_s96_f22_d3_m1_1_1_1/openvino/model.xml\nModel BIN: /home/mount/experiments/exp_1762448723_5984_s96_f22_d3_m1_1_1_1/openvino/model.bin\n[OK] OpenVINO inference engine imported successfully\nDevice: MYRIAD\nModel XML: /home/mount/experiments/exp_1762448723_5984_s96_f22_d3_m1_1_1_1/openvino/model.xml\nModel BIN: /home/mount/experiments/exp_1762448723_5984_s96_f22_d3_m1_1_1_1/openvino/model.bin\nInput file: /home/mount/experiments/exp_1762448723_5984_s96_f22_d3_m1_1_1_1/onnx/sample_input.npy\nInitializing OpenVINO Runtime Core...\nAvailable devices:\n[E:] [BSL] found 0 ioexpander device\n ['CPU', 'GNA', 'MYRIAD']\nLoading network...\nInput blob: input\nInput shape: [1, 8, 96, 96]\nOutput blob: Conv_104\nOutput shape: [1, 3, 96, 96]\nLoading network to MYRIAD...\nLoading input data...\nInput data shape: (1, 8, 96, 96)\nInput data type: float32\nRunning inference...\n[OK] Inference completed!\nInference time: 0.175996 seconds\nOutput shape: (1, 3, 96, 96)\nOutput dtype: float32\nOutput range: [-0.668945, 0.857422]\nMyriad inference completed!\n",
135
+ "stderr": ""
136
+ },
137
+ "end_time": "2025-11-06T18:05:30.696806"
138
+ }
exp_1762448723_5984_s96_f22_d3_m1_1_1_1/run_inference.sh ADDED
@@ -0,0 +1,19 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #!/bin/bash
2
+ # Source OpenVINO environment
3
+ source /opt/intel/openvino/bin/setupvars.sh
4
+
5
+ PYPATH=/home/mount/scripts/simple_inference.py
6
+
7
+ echo "Starting Myriad inference..."
8
+ echo "Input: /home/mount/experiments/exp_1762448723_5984_s96_f22_d3_m1_1_1_1/onnx/sample_input.npy"
9
+ echo "Model XML: /home/mount/experiments/exp_1762448723_5984_s96_f22_d3_m1_1_1_1/openvino/model.xml"
10
+ echo "Model BIN: /home/mount/experiments/exp_1762448723_5984_s96_f22_d3_m1_1_1_1/openvino/model.bin"
11
+
12
+ python3 $PYPATH \
13
+ --input_filepath /home/mount/experiments/exp_1762448723_5984_s96_f22_d3_m1_1_1_1/onnx/sample_input.npy \
14
+ --device_name MYRIAD \
15
+ --model_bin /home/mount/experiments/exp_1762448723_5984_s96_f22_d3_m1_1_1_1/openvino/model.bin \
16
+ --model_xml /home/mount/experiments/exp_1762448723_5984_s96_f22_d3_m1_1_1_1/openvino/model.xml \
17
+ --output /home/mount/experiments/exp_1762448723_5984_s96_f22_d3_m1_1_1_1/inference/
18
+
19
+ echo "Myriad inference completed!"
exp_1762448809_7498_s96_f24_d3_m2_2_2_2/convert_openvino.sh ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #!/bin/bash
2
+ # Source OpenVINO environment
3
+ source /opt/intel/openvino/bin/setupvars.sh
4
+
5
+ # OpenVINO Model Optimizer script for ONNX model
6
+ PYPATH=/opt/intel/openvino_2020.3.194/deployment_tools/model_optimizer/mo.py
7
+ ONNX_MODEL=/home/mount/experiments/exp_1762448809_7498_s96_f24_d3_m2_2_2_2/onnx/model.onnx
8
+ OUTPUT_DIR=/home/mount/experiments/exp_1762448809_7498_s96_f24_d3_m2_2_2_2/openvino
9
+
10
+ echo "Converting ONNX model to OpenVINO IR format..."
11
+ echo "Input model: $ONNX_MODEL"
12
+ echo "Output directory: $OUTPUT_DIR"
13
+
14
+ # Create output directory if it doesn't exist
15
+ mkdir -p "$OUTPUT_DIR"
16
+
17
+ python3 $PYPATH \
18
+ --input_model "$ONNX_MODEL" \
19
+ --data_type FP16 \
20
+ --input_shape "[1,8,96,96]" \
21
+ --mean_values "[0,0,0,0,0,0,0,0]" \
22
+ --scale_values "[1,1,1,1,1,1,1,1]" \
23
+ --progress \
24
+ --stream_output \
25
+ --output_dir "$OUTPUT_DIR" \
26
+ --model_name model
27
+
28
+ echo "OpenVINO conversion completed!"
29
+ echo "Generated files in $OUTPUT_DIR:"
30
+ ls -la "$OUTPUT_DIR/"
exp_1762448809_7498_s96_f24_d3_m2_2_2_2/experiment.log ADDED
@@ -0,0 +1,98 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2025-11-06 18:06:49 - INFO - === Model Creation Phase ===
2
+ 2025-11-06 18:06:49 - INFO - Creating UNet model with config: {'input_size': 96, 'base_filters': 24, 'depth': 3, 'channel_multipliers': [2, 2, 2, 2], 'n_channels': 8, 'n_classes': 3}
3
+ 2025-11-06 18:06:49 - INFO - Model created successfully:
4
+ 2025-11-06 18:06:49 - INFO - Depth: 3
5
+ 2025-11-06 18:06:49 - INFO - Channels per level: [48, 48, 48, 48]
6
+ 2025-11-06 18:06:49 - INFO - Total parameters: 190,851
7
+ 2025-11-06 18:06:49 - INFO - Parameter memory: 0.73 MB
8
+ 2025-11-06 18:06:49 - INFO - === Detailed Architecture Analysis ===
9
+ 2025-11-06 18:06:49 - INFO - Input vector dimension: 73,728 (8 × 96 × 96)
10
+ 2025-11-06 18:06:49 - INFO - Component-wise parameter breakdown:
11
+ 2025-11-06 18:06:49 - INFO - Encoder - encoders.0.channel_proj: 432 parameters
12
+ 2025-11-06 18:06:49 - INFO - Encoder - encoders.0.convnext_block.dwconv: 2,400 parameters
13
+ 2025-11-06 18:06:49 - INFO - Encoder - encoders.0.convnext_block.norm: 96 parameters
14
+ 2025-11-06 18:06:49 - INFO - Encoder - encoders.0.convnext_block.pwconv1: 9,408 parameters
15
+ 2025-11-06 18:06:49 - INFO - Encoder - encoders.0.convnext_block.pwconv2: 9,264 parameters
16
+ 2025-11-06 18:06:49 - INFO - Encoder - encoders.1.convnext_block.dwconv: 2,400 parameters
17
+ 2025-11-06 18:06:49 - INFO - Encoder - encoders.1.convnext_block.norm: 96 parameters
18
+ 2025-11-06 18:06:49 - INFO - Encoder - encoders.1.convnext_block.pwconv1: 9,408 parameters
19
+ 2025-11-06 18:06:49 - INFO - Encoder - encoders.1.convnext_block.pwconv2: 9,264 parameters
20
+ 2025-11-06 18:06:49 - INFO - Encoder - encoders.2.convnext_block.dwconv: 2,400 parameters
21
+ 2025-11-06 18:06:49 - INFO - Encoder - encoders.2.convnext_block.norm: 96 parameters
22
+ 2025-11-06 18:06:49 - INFO - Encoder - encoders.2.convnext_block.pwconv1: 9,408 parameters
23
+ 2025-11-06 18:06:49 - INFO - Encoder - encoders.2.convnext_block.pwconv2: 9,264 parameters
24
+ 2025-11-06 18:06:49 - INFO - Bottleneck - bottleneck.convnext_block.dwconv: 2,400 parameters
25
+ 2025-11-06 18:06:49 - INFO - Bottleneck - bottleneck.convnext_block.norm: 96 parameters
26
+ 2025-11-06 18:06:49 - INFO - Bottleneck - bottleneck.convnext_block.pwconv1: 9,408 parameters
27
+ 2025-11-06 18:06:49 - INFO - Bottleneck - bottleneck.convnext_block.pwconv2: 9,264 parameters
28
+ 2025-11-06 18:06:49 - INFO - Decoder - upsamplers.0: 9,264 parameters
29
+ 2025-11-06 18:06:49 - INFO - Decoder - upsamplers.1: 9,264 parameters
30
+ 2025-11-06 18:06:49 - INFO - Decoder - upsamplers.2: 9,264 parameters
31
+ 2025-11-06 18:06:49 - INFO - Decoder - decoders.0.channel_proj: 4,656 parameters
32
+ 2025-11-06 18:06:49 - INFO - Decoder - decoders.0.convnext_block.dwconv: 2,400 parameters
33
+ 2025-11-06 18:06:49 - INFO - Decoder - decoders.0.convnext_block.norm: 96 parameters
34
+ 2025-11-06 18:06:49 - INFO - Decoder - decoders.0.convnext_block.pwconv1: 9,408 parameters
35
+ 2025-11-06 18:06:49 - INFO - Decoder - decoders.0.convnext_block.pwconv2: 9,264 parameters
36
+ 2025-11-06 18:06:49 - INFO - Decoder - decoders.1.channel_proj: 4,656 parameters
37
+ 2025-11-06 18:06:49 - INFO - Decoder - decoders.1.convnext_block.dwconv: 2,400 parameters
38
+ 2025-11-06 18:06:49 - INFO - Decoder - decoders.1.convnext_block.norm: 96 parameters
39
+ 2025-11-06 18:06:49 - INFO - Decoder - decoders.1.convnext_block.pwconv1: 9,408 parameters
40
+ 2025-11-06 18:06:49 - INFO - Decoder - decoders.1.convnext_block.pwconv2: 9,264 parameters
41
+ 2025-11-06 18:06:49 - INFO - Decoder - decoders.2.channel_proj: 4,656 parameters
42
+ 2025-11-06 18:06:49 - INFO - Decoder - decoders.2.convnext_block.dwconv: 2,400 parameters
43
+ 2025-11-06 18:06:49 - INFO - Decoder - decoders.2.convnext_block.norm: 96 parameters
44
+ 2025-11-06 18:06:49 - INFO - Decoder - decoders.2.convnext_block.pwconv1: 9,408 parameters
45
+ 2025-11-06 18:06:49 - INFO - Decoder - decoders.2.convnext_block.pwconv2: 9,264 parameters
46
+ 2025-11-06 18:06:49 - INFO - Other - final_conv: 147 parameters
47
+ 2025-11-06 18:06:49 - INFO - Parameter distribution summary:
48
+ 2025-11-06 18:06:49 - INFO - Encoder parameters: 63,936 (33.6%)
49
+ 2025-11-06 18:06:49 - INFO - Decoder parameters: 105,264 (55.3%)
50
+ 2025-11-06 18:06:49 - INFO - Bottleneck parameters: 21,168 (11.1%)
51
+ 2025-11-06 18:06:49 - INFO - Other parameters: 147 (0.1%)
52
+ 2025-11-06 18:06:49 - INFO - Latent space dimensions (feature maps at each level):
53
+ 2025-11-06 18:06:49 - INFO - Level 0: 48 × 96 × 96 = 442,368 elements
54
+ 2025-11-06 18:06:49 - INFO - Level 1: 48 × 48 × 48 = 110,592 elements
55
+ 2025-11-06 18:06:49 - INFO - Level 2: 48 × 24 × 24 = 27,648 elements
56
+ 2025-11-06 18:06:49 - INFO - Level 3: 48 × 12 × 12 = 6,912 elements
57
+ 2025-11-06 18:06:49 - INFO - Skip connection dimensions:
58
+ 2025-11-06 18:06:49 - INFO - Skip 0: 48 × 96 × 96 = 442,368 elements
59
+ 2025-11-06 18:06:49 - INFO - Skip 1: 48 × 48 × 48 = 110,592 elements
60
+ 2025-11-06 18:06:49 - INFO - Skip 2: 48 × 24 × 24 = 27,648 elements
61
+ 2025-11-06 18:06:49 - INFO - Memory analysis:
62
+ 2025-11-06 18:06:49 - INFO - Peak feature map memory (inference): 2.52 MB
63
+ 2025-11-06 18:06:49 - INFO - Peak feature map memory (training): 5.04 MB (with gradients)
64
+ 2025-11-06 18:06:49 - INFO - Output vector dimension: 27,648 (3 × 96 × 96)
65
+ 2025-11-06 18:06:49 - INFO - PyTorch model saved to: /home/philab/Desktop/hydranet/experiments/exp_1762448809_7498_s96_f24_d3_m2_2_2_2/pytorch/model.pt
66
+ 2025-11-06 18:06:49 - INFO - === ONNX Conversion Phase ===
67
+ 2025-11-06 18:06:49 - INFO - === Model Export Diagnostics ===
68
+ 2025-11-06 18:06:49 - INFO - PyTorch version: 1.9.0+cu102
69
+ 2025-11-06 18:06:49 - INFO - Model parameters: 190,851
70
+ 2025-11-06 18:06:49 - INFO - Model memory: 0.73 MB
71
+ 2025-11-06 18:06:49 - INFO - Starting ONNX export with opset version 11
72
+ 2025-11-06 18:06:49 - INFO - Model input shape: torch.Size([1, 8, 96, 96])
73
+ 2025-11-06 18:06:49 - INFO - Model input dtype: torch.float32
74
+ 2025-11-06 18:06:49 - INFO - Forward pass successful. Output shape: torch.Size([1, 3, 96, 96])
75
+ 2025-11-06 18:06:49 - INFO - Output dtype: torch.float32
76
+ 2025-11-06 18:06:49 - INFO - Output value range: [-0.4865, 0.6665]
77
+ 2025-11-06 18:06:49 - INFO - Model successfully exported to /home/philab/Desktop/hydranet/experiments/exp_1762448809_7498_s96_f24_d3_m2_2_2_2/onnx/model.onnx
78
+ 2025-11-06 18:06:49 - INFO - ONNX model size: 0.74 MB
79
+ 2025-11-06 18:06:49 - INFO - Saved dummy input with shape (1, 8, 96, 96) to /home/philab/Desktop/hydranet/experiments/exp_1762448809_7498_s96_f24_d3_m2_2_2_2/onnx/sample_input.npy
80
+ 2025-11-06 18:06:49 - INFO - Input data type: float32
81
+ 2025-11-06 18:06:49 - INFO - Input value range: [-4.5734, 4.4761]
82
+ 2025-11-06 18:06:49 - INFO - === OpenVINO Conversion Phase ===
83
+ 2025-11-06 18:06:49 - INFO - Starting OpenVINO conversion in Docker container...
84
+ 2025-11-06 18:06:53 - INFO - OpenVINO conversion completed in 4.00 seconds
85
+ 2025-11-06 18:06:53 - INFO - OpenVINO model files created:
86
+ 2025-11-06 18:06:53 - INFO - XML file: /home/philab/Desktop/hydranet/experiments/exp_1762448809_7498_s96_f24_d3_m2_2_2_2/openvino/model.xml (0.08 MB)
87
+ 2025-11-06 18:06:53 - INFO - BIN file: /home/philab/Desktop/hydranet/experiments/exp_1762448809_7498_s96_f24_d3_m2_2_2_2/openvino/model.bin (0.36 MB)
88
+ 2025-11-06 18:06:53 - INFO - === Myriad Inference Phase ===
89
+ 2025-11-06 18:06:53 - INFO - Starting Myriad inference in Docker container...
90
+ 2025-11-06 18:06:56 - INFO - Myriad inference completed in 2.77 seconds
91
+ 2025-11-06 18:06:56 - INFO - Actual inference time: 0.171749 seconds
92
+ 2025-11-06 18:06:56 - INFO - ✅ Complete pipeline executed successfully!
93
+ 2025-11-06 18:06:56 - INFO - ✅ Experiment 46 completed successfully
94
+ 2025-11-06 18:06:56 - INFO - Inference time: 0.171749s
95
+ 2025-11-06 18:06:56 - INFO -
96
+ === Experiment 47/2475 ===
97
+ 2025-11-06 18:06:56 - INFO - Experiment ID: exp_1762448816_2408_s96_f24_d3_m1_2_2_2
98
+ 2025-11-06 18:06:56 - INFO - Experiment directory: /home/philab/Desktop/hydranet/experiments/exp_1762448816_2408_s96_f24_d3_m1_2_2_2
exp_1762448809_7498_s96_f24_d3_m2_2_2_2/model_info.json ADDED
@@ -0,0 +1,138 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "config": {
3
+ "input_size": 96,
4
+ "base_filters": 24,
5
+ "depth": 3,
6
+ "channel_multipliers": [
7
+ 2,
8
+ 2,
9
+ 2,
10
+ 2
11
+ ],
12
+ "n_channels": 8,
13
+ "n_classes": 3
14
+ },
15
+ "model_info": {
16
+ "depth": 3,
17
+ "channels_per_level": [
18
+ 48,
19
+ 48,
20
+ 48,
21
+ 48
22
+ ],
23
+ "channel_multipliers": [
24
+ 2,
25
+ 2,
26
+ 2,
27
+ 2
28
+ ],
29
+ "total_parameters": 190851,
30
+ "trainable_parameters": 190851,
31
+ "model_size_mb": 0.7280387878417969
32
+ },
33
+ "architecture_stats": {
34
+ "input_dimension": 73728,
35
+ "output_dimension": 27648,
36
+ "parameter_distribution": {
37
+ "encoder_params": 63936,
38
+ "decoder_params": 105264,
39
+ "bottleneck_params": 21168,
40
+ "other_params": 147,
41
+ "total_params": 190515,
42
+ "encoder_percentage": 33.55956223919377,
43
+ "decoder_percentage": 55.252342335249196,
44
+ "bottleneck_percentage": 11.110936146760098,
45
+ "other_percentage": 0.07715927879694512
46
+ },
47
+ "component_breakdown": {
48
+ "Encoder - encoders.0.channel_proj": 432,
49
+ "Encoder - encoders.0.convnext_block.dwconv": 2400,
50
+ "Encoder - encoders.0.convnext_block.norm": 96,
51
+ "Encoder - encoders.0.convnext_block.pwconv1": 9408,
52
+ "Encoder - encoders.0.convnext_block.pwconv2": 9264,
53
+ "Encoder - encoders.1.convnext_block.dwconv": 2400,
54
+ "Encoder - encoders.1.convnext_block.norm": 96,
55
+ "Encoder - encoders.1.convnext_block.pwconv1": 9408,
56
+ "Encoder - encoders.1.convnext_block.pwconv2": 9264,
57
+ "Encoder - encoders.2.convnext_block.dwconv": 2400,
58
+ "Encoder - encoders.2.convnext_block.norm": 96,
59
+ "Encoder - encoders.2.convnext_block.pwconv1": 9408,
60
+ "Encoder - encoders.2.convnext_block.pwconv2": 9264,
61
+ "Bottleneck - bottleneck.convnext_block.dwconv": 2400,
62
+ "Bottleneck - bottleneck.convnext_block.norm": 96,
63
+ "Bottleneck - bottleneck.convnext_block.pwconv1": 9408,
64
+ "Bottleneck - bottleneck.convnext_block.pwconv2": 9264,
65
+ "Decoder - upsamplers.0": 9264,
66
+ "Decoder - upsamplers.1": 9264,
67
+ "Decoder - upsamplers.2": 9264,
68
+ "Decoder - decoders.0.channel_proj": 4656,
69
+ "Decoder - decoders.0.convnext_block.dwconv": 2400,
70
+ "Decoder - decoders.0.convnext_block.norm": 96,
71
+ "Decoder - decoders.0.convnext_block.pwconv1": 9408,
72
+ "Decoder - decoders.0.convnext_block.pwconv2": 9264,
73
+ "Decoder - decoders.1.channel_proj": 4656,
74
+ "Decoder - decoders.1.convnext_block.dwconv": 2400,
75
+ "Decoder - decoders.1.convnext_block.norm": 96,
76
+ "Decoder - decoders.1.convnext_block.pwconv1": 9408,
77
+ "Decoder - decoders.1.convnext_block.pwconv2": 9264,
78
+ "Decoder - decoders.2.channel_proj": 4656,
79
+ "Decoder - decoders.2.convnext_block.dwconv": 2400,
80
+ "Decoder - decoders.2.convnext_block.norm": 96,
81
+ "Decoder - decoders.2.convnext_block.pwconv1": 9408,
82
+ "Decoder - decoders.2.convnext_block.pwconv2": 9264,
83
+ "Other - final_conv": 147
84
+ },
85
+ "latent_dimensions": {
86
+ "Level_0": {
87
+ "channels": 48,
88
+ "height": 96,
89
+ "width": 96,
90
+ "total_elements": 442368
91
+ },
92
+ "Level_1": {
93
+ "channels": 48,
94
+ "height": 48,
95
+ "width": 48,
96
+ "total_elements": 110592
97
+ },
98
+ "Level_2": {
99
+ "channels": 48,
100
+ "height": 24,
101
+ "width": 24,
102
+ "total_elements": 27648
103
+ },
104
+ "Level_3": {
105
+ "channels": 48,
106
+ "height": 12,
107
+ "width": 12,
108
+ "total_elements": 6912
109
+ }
110
+ },
111
+ "skip_dimensions": {
112
+ "Skip_0": {
113
+ "channels": 48,
114
+ "height": 96,
115
+ "width": 96,
116
+ "total_elements": 442368
117
+ },
118
+ "Skip_1": {
119
+ "channels": 48,
120
+ "height": 48,
121
+ "width": 48,
122
+ "total_elements": 110592
123
+ },
124
+ "Skip_2": {
125
+ "channels": 48,
126
+ "height": 24,
127
+ "width": 24,
128
+ "total_elements": 27648
129
+ }
130
+ },
131
+ "memory_analysis": {
132
+ "peak_memory_inference_mb": 2.5224609375,
133
+ "peak_memory_training_mb": 5.044921875,
134
+ "peak_elements": 661248
135
+ }
136
+ },
137
+ "timestamp": "2025-11-06T18:06:49.610394"
138
+ }
exp_1762448809_7498_s96_f24_d3_m2_2_2_2/pipeline_results.json ADDED
@@ -0,0 +1,138 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "experiment_id": "exp_1762448809_5070_s96_f24_d3_m2_2_2_2",
3
+ "config": {
4
+ "input_size": 96,
5
+ "base_filters": 24,
6
+ "depth": 3,
7
+ "channel_multipliers": [
8
+ 2,
9
+ 2,
10
+ 2,
11
+ 2
12
+ ],
13
+ "n_channels": 8,
14
+ "n_classes": 3
15
+ },
16
+ "start_time": "2025-11-06T18:06:49.604051",
17
+ "success": true,
18
+ "timings": {
19
+ "onnx_conversion": 0.2828185979742557,
20
+ "openvino_conversion": 3.9952829838730395,
21
+ "inference_total": 2.769099799916148,
22
+ "inference_actual": 0.171749,
23
+ "total": 7.058676311047748
24
+ },
25
+ "errors": [],
26
+ "architecture_stats": {
27
+ "input_dimension": 73728,
28
+ "output_dimension": 27648,
29
+ "parameter_distribution": {
30
+ "encoder_params": 63936,
31
+ "decoder_params": 105264,
32
+ "bottleneck_params": 21168,
33
+ "other_params": 147,
34
+ "total_params": 190515,
35
+ "encoder_percentage": 33.55956223919377,
36
+ "decoder_percentage": 55.252342335249196,
37
+ "bottleneck_percentage": 11.110936146760098,
38
+ "other_percentage": 0.07715927879694512
39
+ },
40
+ "component_breakdown": {
41
+ "Encoder - encoders.0.channel_proj": 432,
42
+ "Encoder - encoders.0.convnext_block.dwconv": 2400,
43
+ "Encoder - encoders.0.convnext_block.norm": 96,
44
+ "Encoder - encoders.0.convnext_block.pwconv1": 9408,
45
+ "Encoder - encoders.0.convnext_block.pwconv2": 9264,
46
+ "Encoder - encoders.1.convnext_block.dwconv": 2400,
47
+ "Encoder - encoders.1.convnext_block.norm": 96,
48
+ "Encoder - encoders.1.convnext_block.pwconv1": 9408,
49
+ "Encoder - encoders.1.convnext_block.pwconv2": 9264,
50
+ "Encoder - encoders.2.convnext_block.dwconv": 2400,
51
+ "Encoder - encoders.2.convnext_block.norm": 96,
52
+ "Encoder - encoders.2.convnext_block.pwconv1": 9408,
53
+ "Encoder - encoders.2.convnext_block.pwconv2": 9264,
54
+ "Bottleneck - bottleneck.convnext_block.dwconv": 2400,
55
+ "Bottleneck - bottleneck.convnext_block.norm": 96,
56
+ "Bottleneck - bottleneck.convnext_block.pwconv1": 9408,
57
+ "Bottleneck - bottleneck.convnext_block.pwconv2": 9264,
58
+ "Decoder - upsamplers.0": 9264,
59
+ "Decoder - upsamplers.1": 9264,
60
+ "Decoder - upsamplers.2": 9264,
61
+ "Decoder - decoders.0.channel_proj": 4656,
62
+ "Decoder - decoders.0.convnext_block.dwconv": 2400,
63
+ "Decoder - decoders.0.convnext_block.norm": 96,
64
+ "Decoder - decoders.0.convnext_block.pwconv1": 9408,
65
+ "Decoder - decoders.0.convnext_block.pwconv2": 9264,
66
+ "Decoder - decoders.1.channel_proj": 4656,
67
+ "Decoder - decoders.1.convnext_block.dwconv": 2400,
68
+ "Decoder - decoders.1.convnext_block.norm": 96,
69
+ "Decoder - decoders.1.convnext_block.pwconv1": 9408,
70
+ "Decoder - decoders.1.convnext_block.pwconv2": 9264,
71
+ "Decoder - decoders.2.channel_proj": 4656,
72
+ "Decoder - decoders.2.convnext_block.dwconv": 2400,
73
+ "Decoder - decoders.2.convnext_block.norm": 96,
74
+ "Decoder - decoders.2.convnext_block.pwconv1": 9408,
75
+ "Decoder - decoders.2.convnext_block.pwconv2": 9264,
76
+ "Other - final_conv": 147
77
+ },
78
+ "latent_dimensions": {
79
+ "Level_0": {
80
+ "channels": 48,
81
+ "height": 96,
82
+ "width": 96,
83
+ "total_elements": 442368
84
+ },
85
+ "Level_1": {
86
+ "channels": 48,
87
+ "height": 48,
88
+ "width": 48,
89
+ "total_elements": 110592
90
+ },
91
+ "Level_2": {
92
+ "channels": 48,
93
+ "height": 24,
94
+ "width": 24,
95
+ "total_elements": 27648
96
+ },
97
+ "Level_3": {
98
+ "channels": 48,
99
+ "height": 12,
100
+ "width": 12,
101
+ "total_elements": 6912
102
+ }
103
+ },
104
+ "skip_dimensions": {
105
+ "Skip_0": {
106
+ "channels": 48,
107
+ "height": 96,
108
+ "width": 96,
109
+ "total_elements": 442368
110
+ },
111
+ "Skip_1": {
112
+ "channels": 48,
113
+ "height": 48,
114
+ "width": 48,
115
+ "total_elements": 110592
116
+ },
117
+ "Skip_2": {
118
+ "channels": 48,
119
+ "height": 24,
120
+ "width": 24,
121
+ "total_elements": 27648
122
+ }
123
+ },
124
+ "memory_analysis": {
125
+ "peak_memory_inference_mb": 2.5224609375,
126
+ "peak_memory_training_mb": 5.044921875,
127
+ "peak_elements": 661248
128
+ }
129
+ },
130
+ "inference_results": {
131
+ "success": true,
132
+ "total_time": 2.769099799916148,
133
+ "inference_time": 0.171749,
134
+ "stdout": "[setupvars.sh] OpenVINO environment initialized\nStarting Myriad inference...\nInput: /home/mount/experiments/exp_1762448809_7498_s96_f24_d3_m2_2_2_2/onnx/sample_input.npy\nModel XML: /home/mount/experiments/exp_1762448809_7498_s96_f24_d3_m2_2_2_2/openvino/model.xml\nModel BIN: /home/mount/experiments/exp_1762448809_7498_s96_f24_d3_m2_2_2_2/openvino/model.bin\n[OK] OpenVINO inference engine imported successfully\nDevice: MYRIAD\nModel XML: /home/mount/experiments/exp_1762448809_7498_s96_f24_d3_m2_2_2_2/openvino/model.xml\nModel BIN: /home/mount/experiments/exp_1762448809_7498_s96_f24_d3_m2_2_2_2/openvino/model.bin\nInput file: /home/mount/experiments/exp_1762448809_7498_s96_f24_d3_m2_2_2_2/onnx/sample_input.npy\nInitializing OpenVINO Runtime Core...\nAvailable devices:\n[E:] [BSL] found 0 ioexpander device\n ['CPU', 'GNA', 'MYRIAD']\nLoading network...\nInput blob: input\nInput shape: [1, 8, 96, 96]\nOutput blob: Conv_104\nOutput shape: [1, 3, 96, 96]\nLoading network to MYRIAD...\nLoading input data...\nInput data shape: (1, 8, 96, 96)\nInput data type: float32\nRunning inference...\n[OK] Inference completed!\nInference time: 0.171749 seconds\nOutput shape: (1, 3, 96, 96)\nOutput dtype: float32\nOutput range: [-0.486084, 0.666016]\nMyriad inference completed!\n",
135
+ "stderr": ""
136
+ },
137
+ "end_time": "2025-11-06T18:06:56.662717"
138
+ }
exp_1762448809_7498_s96_f24_d3_m2_2_2_2/run_inference.sh ADDED
@@ -0,0 +1,19 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #!/bin/bash
2
+ # Source OpenVINO environment
3
+ source /opt/intel/openvino/bin/setupvars.sh
4
+
5
+ PYPATH=/home/mount/scripts/simple_inference.py
6
+
7
+ echo "Starting Myriad inference..."
8
+ echo "Input: /home/mount/experiments/exp_1762448809_7498_s96_f24_d3_m2_2_2_2/onnx/sample_input.npy"
9
+ echo "Model XML: /home/mount/experiments/exp_1762448809_7498_s96_f24_d3_m2_2_2_2/openvino/model.xml"
10
+ echo "Model BIN: /home/mount/experiments/exp_1762448809_7498_s96_f24_d3_m2_2_2_2/openvino/model.bin"
11
+
12
+ python3 $PYPATH \
13
+ --input_filepath /home/mount/experiments/exp_1762448809_7498_s96_f24_d3_m2_2_2_2/onnx/sample_input.npy \
14
+ --device_name MYRIAD \
15
+ --model_bin /home/mount/experiments/exp_1762448809_7498_s96_f24_d3_m2_2_2_2/openvino/model.bin \
16
+ --model_xml /home/mount/experiments/exp_1762448809_7498_s96_f24_d3_m2_2_2_2/openvino/model.xml \
17
+ --output /home/mount/experiments/exp_1762448809_7498_s96_f24_d3_m2_2_2_2/inference/
18
+
19
+ echo "Myriad inference completed!"
exp_1762448858_8082_s96_f24_d3_m1_2_3_4/convert_openvino.sh ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #!/bin/bash
2
+ # Source OpenVINO environment
3
+ source /opt/intel/openvino/bin/setupvars.sh
4
+
5
+ # OpenVINO Model Optimizer script for ONNX model
6
+ PYPATH=/opt/intel/openvino_2020.3.194/deployment_tools/model_optimizer/mo.py
7
+ ONNX_MODEL=/home/mount/experiments/exp_1762448858_8082_s96_f24_d3_m1_2_3_4/onnx/model.onnx
8
+ OUTPUT_DIR=/home/mount/experiments/exp_1762448858_8082_s96_f24_d3_m1_2_3_4/openvino
9
+
10
+ echo "Converting ONNX model to OpenVINO IR format..."
11
+ echo "Input model: $ONNX_MODEL"
12
+ echo "Output directory: $OUTPUT_DIR"
13
+
14
+ # Create output directory if it doesn't exist
15
+ mkdir -p "$OUTPUT_DIR"
16
+
17
+ python3 $PYPATH \
18
+ --input_model "$ONNX_MODEL" \
19
+ --data_type FP16 \
20
+ --input_shape "[1,8,96,96]" \
21
+ --mean_values "[0,0,0,0,0,0,0,0]" \
22
+ --scale_values "[1,1,1,1,1,1,1,1]" \
23
+ --progress \
24
+ --stream_output \
25
+ --output_dir "$OUTPUT_DIR" \
26
+ --model_name model
27
+
28
+ echo "OpenVINO conversion completed!"
29
+ echo "Generated files in $OUTPUT_DIR:"
30
+ ls -la "$OUTPUT_DIR/"
exp_1762448858_8082_s96_f24_d3_m1_2_3_4/experiment.log ADDED
@@ -0,0 +1,101 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2025-11-06 18:07:38 - INFO - === Model Creation Phase ===
2
+ 2025-11-06 18:07:38 - INFO - Creating UNet model with config: {'input_size': 96, 'base_filters': 24, 'depth': 3, 'channel_multipliers': [1, 2, 3, 4], 'n_channels': 8, 'n_classes': 3}
3
+ 2025-11-06 18:07:38 - INFO - Model created successfully:
4
+ 2025-11-06 18:07:38 - INFO - Depth: 3
5
+ 2025-11-06 18:07:38 - INFO - Channels per level: [24, 48, 72, 96]
6
+ 2025-11-06 18:07:38 - INFO - Total parameters: 323,811
7
+ 2025-11-06 18:07:38 - INFO - Parameter memory: 1.24 MB
8
+ 2025-11-06 18:07:38 - INFO - === Detailed Architecture Analysis ===
9
+ 2025-11-06 18:07:38 - INFO - Input vector dimension: 73,728 (8 × 96 × 96)
10
+ 2025-11-06 18:07:38 - INFO - Component-wise parameter breakdown:
11
+ 2025-11-06 18:07:38 - INFO - Encoder - encoders.0.channel_proj: 216 parameters
12
+ 2025-11-06 18:07:38 - INFO - Encoder - encoders.0.convnext_block.dwconv: 1,200 parameters
13
+ 2025-11-06 18:07:38 - INFO - Encoder - encoders.0.convnext_block.norm: 48 parameters
14
+ 2025-11-06 18:07:38 - INFO - Encoder - encoders.0.convnext_block.pwconv1: 2,400 parameters
15
+ 2025-11-06 18:07:38 - INFO - Encoder - encoders.0.convnext_block.pwconv2: 2,328 parameters
16
+ 2025-11-06 18:07:38 - INFO - Encoder - encoders.1.channel_proj: 1,200 parameters
17
+ 2025-11-06 18:07:38 - INFO - Encoder - encoders.1.convnext_block.dwconv: 2,400 parameters
18
+ 2025-11-06 18:07:38 - INFO - Encoder - encoders.1.convnext_block.norm: 96 parameters
19
+ 2025-11-06 18:07:38 - INFO - Encoder - encoders.1.convnext_block.pwconv1: 9,408 parameters
20
+ 2025-11-06 18:07:38 - INFO - Encoder - encoders.1.convnext_block.pwconv2: 9,264 parameters
21
+ 2025-11-06 18:07:38 - INFO - Encoder - encoders.2.channel_proj: 3,528 parameters
22
+ 2025-11-06 18:07:38 - INFO - Encoder - encoders.2.convnext_block.dwconv: 3,600 parameters
23
+ 2025-11-06 18:07:38 - INFO - Encoder - encoders.2.convnext_block.norm: 144 parameters
24
+ 2025-11-06 18:07:38 - INFO - Encoder - encoders.2.convnext_block.pwconv1: 21,024 parameters
25
+ 2025-11-06 18:07:38 - INFO - Encoder - encoders.2.convnext_block.pwconv2: 20,808 parameters
26
+ 2025-11-06 18:07:38 - INFO - Bottleneck - bottleneck.channel_proj: 7,008 parameters
27
+ 2025-11-06 18:07:38 - INFO - Bottleneck - bottleneck.convnext_block.dwconv: 4,800 parameters
28
+ 2025-11-06 18:07:38 - INFO - Bottleneck - bottleneck.convnext_block.norm: 192 parameters
29
+ 2025-11-06 18:07:38 - INFO - Bottleneck - bottleneck.convnext_block.pwconv1: 37,248 parameters
30
+ 2025-11-06 18:07:38 - INFO - Bottleneck - bottleneck.convnext_block.pwconv2: 36,960 parameters
31
+ 2025-11-06 18:07:38 - INFO - Decoder - upsamplers.0: 36,960 parameters
32
+ 2025-11-06 18:07:38 - INFO - Decoder - upsamplers.1: 20,808 parameters
33
+ 2025-11-06 18:07:38 - INFO - Decoder - upsamplers.2: 9,264 parameters
34
+ 2025-11-06 18:07:38 - INFO - Decoder - decoders.0.channel_proj: 12,168 parameters
35
+ 2025-11-06 18:07:38 - INFO - Decoder - decoders.0.convnext_block.dwconv: 3,600 parameters
36
+ 2025-11-06 18:07:38 - INFO - Decoder - decoders.0.convnext_block.norm: 144 parameters
37
+ 2025-11-06 18:07:38 - INFO - Decoder - decoders.0.convnext_block.pwconv1: 21,024 parameters
38
+ 2025-11-06 18:07:38 - INFO - Decoder - decoders.0.convnext_block.pwconv2: 20,808 parameters
39
+ 2025-11-06 18:07:38 - INFO - Decoder - decoders.1.channel_proj: 5,808 parameters
40
+ 2025-11-06 18:07:38 - INFO - Decoder - decoders.1.convnext_block.dwconv: 2,400 parameters
41
+ 2025-11-06 18:07:38 - INFO - Decoder - decoders.1.convnext_block.norm: 96 parameters
42
+ 2025-11-06 18:07:38 - INFO - Decoder - decoders.1.convnext_block.pwconv1: 9,408 parameters
43
+ 2025-11-06 18:07:38 - INFO - Decoder - decoders.1.convnext_block.pwconv2: 9,264 parameters
44
+ 2025-11-06 18:07:38 - INFO - Decoder - decoders.2.channel_proj: 1,752 parameters
45
+ 2025-11-06 18:07:38 - INFO - Decoder - decoders.2.convnext_block.dwconv: 1,200 parameters
46
+ 2025-11-06 18:07:38 - INFO - Decoder - decoders.2.convnext_block.norm: 48 parameters
47
+ 2025-11-06 18:07:38 - INFO - Decoder - decoders.2.convnext_block.pwconv1: 2,400 parameters
48
+ 2025-11-06 18:07:38 - INFO - Decoder - decoders.2.convnext_block.pwconv2: 2,328 parameters
49
+ 2025-11-06 18:07:38 - INFO - Other - final_conv: 75 parameters
50
+ 2025-11-06 18:07:38 - INFO - Parameter distribution summary:
51
+ 2025-11-06 18:07:38 - INFO - Encoder parameters: 77,664 (24.0%)
52
+ 2025-11-06 18:07:38 - INFO - Decoder parameters: 159,480 (49.3%)
53
+ 2025-11-06 18:07:38 - INFO - Bottleneck parameters: 86,208 (26.7%)
54
+ 2025-11-06 18:07:38 - INFO - Other parameters: 75 (0.0%)
55
+ 2025-11-06 18:07:38 - INFO - Latent space dimensions (feature maps at each level):
56
+ 2025-11-06 18:07:38 - INFO - Level 0: 24 × 96 × 96 = 221,184 elements
57
+ 2025-11-06 18:07:38 - INFO - Level 1: 48 × 48 × 48 = 110,592 elements
58
+ 2025-11-06 18:07:38 - INFO - Level 2: 72 × 24 × 24 = 41,472 elements
59
+ 2025-11-06 18:07:38 - INFO - Level 3: 96 × 12 × 12 = 13,824 elements
60
+ 2025-11-06 18:07:38 - INFO - Skip connection dimensions:
61
+ 2025-11-06 18:07:38 - INFO - Skip 0: 24 × 96 × 96 = 221,184 elements
62
+ 2025-11-06 18:07:38 - INFO - Skip 1: 48 × 48 × 48 = 110,592 elements
63
+ 2025-11-06 18:07:38 - INFO - Skip 2: 72 × 24 × 24 = 41,472 elements
64
+ 2025-11-06 18:07:38 - INFO - Memory analysis:
65
+ 2025-11-06 18:07:38 - INFO - Peak feature map memory (inference): 1.76 MB
66
+ 2025-11-06 18:07:38 - INFO - Peak feature map memory (training): 3.52 MB (with gradients)
67
+ 2025-11-06 18:07:38 - INFO - Output vector dimension: 27,648 (3 × 96 × 96)
68
+ 2025-11-06 18:07:38 - INFO - PyTorch model saved to: /home/philab/Desktop/hydranet/experiments/exp_1762448858_8082_s96_f24_d3_m1_2_3_4/pytorch/model.pt
69
+ 2025-11-06 18:07:38 - INFO - === ONNX Conversion Phase ===
70
+ 2025-11-06 18:07:38 - INFO - === Model Export Diagnostics ===
71
+ 2025-11-06 18:07:38 - INFO - PyTorch version: 1.9.0+cu102
72
+ 2025-11-06 18:07:38 - INFO - Model parameters: 323,811
73
+ 2025-11-06 18:07:38 - INFO - Model memory: 1.24 MB
74
+ 2025-11-06 18:07:38 - INFO - Starting ONNX export with opset version 11
75
+ 2025-11-06 18:07:38 - INFO - Model input shape: torch.Size([1, 8, 96, 96])
76
+ 2025-11-06 18:07:38 - INFO - Model input dtype: torch.float32
77
+ 2025-11-06 18:07:38 - INFO - Forward pass successful. Output shape: torch.Size([1, 3, 96, 96])
78
+ 2025-11-06 18:07:38 - INFO - Output dtype: torch.float32
79
+ 2025-11-06 18:07:38 - INFO - Output value range: [-0.4544, 0.5538]
80
+ 2025-11-06 18:07:39 - INFO - Model successfully exported to /home/philab/Desktop/hydranet/experiments/exp_1762448858_8082_s96_f24_d3_m1_2_3_4/onnx/model.onnx
81
+ 2025-11-06 18:07:39 - INFO - ONNX model size: 1.24 MB
82
+ 2025-11-06 18:07:39 - INFO - Saved dummy input with shape (1, 8, 96, 96) to /home/philab/Desktop/hydranet/experiments/exp_1762448858_8082_s96_f24_d3_m1_2_3_4/onnx/sample_input.npy
83
+ 2025-11-06 18:07:39 - INFO - Input data type: float32
84
+ 2025-11-06 18:07:39 - INFO - Input value range: [-4.4298, 4.3074]
85
+ 2025-11-06 18:07:39 - INFO - === OpenVINO Conversion Phase ===
86
+ 2025-11-06 18:07:39 - INFO - Starting OpenVINO conversion in Docker container...
87
+ 2025-11-06 18:07:43 - INFO - OpenVINO conversion completed in 4.16 seconds
88
+ 2025-11-06 18:07:43 - INFO - OpenVINO model files created:
89
+ 2025-11-06 18:07:43 - INFO - XML file: /home/philab/Desktop/hydranet/experiments/exp_1762448858_8082_s96_f24_d3_m1_2_3_4/openvino/model.xml (0.09 MB)
90
+ 2025-11-06 18:07:43 - INFO - BIN file: /home/philab/Desktop/hydranet/experiments/exp_1762448858_8082_s96_f24_d3_m1_2_3_4/openvino/model.bin (0.62 MB)
91
+ 2025-11-06 18:07:43 - INFO - === Myriad Inference Phase ===
92
+ 2025-11-06 18:07:43 - INFO - Starting Myriad inference in Docker container...
93
+ 2025-11-06 18:07:46 - INFO - Myriad inference completed in 2.71 seconds
94
+ 2025-11-06 18:07:46 - INFO - Actual inference time: 0.123546 seconds
95
+ 2025-11-06 18:07:46 - INFO - ✅ Complete pipeline executed successfully!
96
+ 2025-11-06 18:07:46 - INFO - ✅ Experiment 53 completed successfully
97
+ 2025-11-06 18:07:46 - INFO - Inference time: 0.123546s
98
+ 2025-11-06 18:07:46 - INFO -
99
+ === Experiment 54/2475 ===
100
+ 2025-11-06 18:07:46 - INFO - Experiment ID: exp_1762448866_1287_s96_f24_d3_m1_3_3_4
101
+ 2025-11-06 18:07:46 - INFO - Experiment directory: /home/philab/Desktop/hydranet/experiments/exp_1762448866_1287_s96_f24_d3_m1_3_3_4
exp_1762448858_8082_s96_f24_d3_m1_2_3_4/model_info.json ADDED
@@ -0,0 +1,141 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "config": {
3
+ "input_size": 96,
4
+ "base_filters": 24,
5
+ "depth": 3,
6
+ "channel_multipliers": [
7
+ 1,
8
+ 2,
9
+ 3,
10
+ 4
11
+ ],
12
+ "n_channels": 8,
13
+ "n_classes": 3
14
+ },
15
+ "model_info": {
16
+ "depth": 3,
17
+ "channels_per_level": [
18
+ 24,
19
+ 48,
20
+ 72,
21
+ 96
22
+ ],
23
+ "channel_multipliers": [
24
+ 1,
25
+ 2,
26
+ 3,
27
+ 4
28
+ ],
29
+ "total_parameters": 323811,
30
+ "trainable_parameters": 323811,
31
+ "model_size_mb": 1.2352409362792969
32
+ },
33
+ "architecture_stats": {
34
+ "input_dimension": 73728,
35
+ "output_dimension": 27648,
36
+ "parameter_distribution": {
37
+ "encoder_params": 77664,
38
+ "decoder_params": 159480,
39
+ "bottleneck_params": 86208,
40
+ "other_params": 75,
41
+ "total_params": 323427,
42
+ "encoder_percentage": 24.01283751820349,
43
+ "decoder_percentage": 49.30942685675593,
44
+ "bottleneck_percentage": 26.654546466436013,
45
+ "other_percentage": 0.02318915860456919
46
+ },
47
+ "component_breakdown": {
48
+ "Encoder - encoders.0.channel_proj": 216,
49
+ "Encoder - encoders.0.convnext_block.dwconv": 1200,
50
+ "Encoder - encoders.0.convnext_block.norm": 48,
51
+ "Encoder - encoders.0.convnext_block.pwconv1": 2400,
52
+ "Encoder - encoders.0.convnext_block.pwconv2": 2328,
53
+ "Encoder - encoders.1.channel_proj": 1200,
54
+ "Encoder - encoders.1.convnext_block.dwconv": 2400,
55
+ "Encoder - encoders.1.convnext_block.norm": 96,
56
+ "Encoder - encoders.1.convnext_block.pwconv1": 9408,
57
+ "Encoder - encoders.1.convnext_block.pwconv2": 9264,
58
+ "Encoder - encoders.2.channel_proj": 3528,
59
+ "Encoder - encoders.2.convnext_block.dwconv": 3600,
60
+ "Encoder - encoders.2.convnext_block.norm": 144,
61
+ "Encoder - encoders.2.convnext_block.pwconv1": 21024,
62
+ "Encoder - encoders.2.convnext_block.pwconv2": 20808,
63
+ "Bottleneck - bottleneck.channel_proj": 7008,
64
+ "Bottleneck - bottleneck.convnext_block.dwconv": 4800,
65
+ "Bottleneck - bottleneck.convnext_block.norm": 192,
66
+ "Bottleneck - bottleneck.convnext_block.pwconv1": 37248,
67
+ "Bottleneck - bottleneck.convnext_block.pwconv2": 36960,
68
+ "Decoder - upsamplers.0": 36960,
69
+ "Decoder - upsamplers.1": 20808,
70
+ "Decoder - upsamplers.2": 9264,
71
+ "Decoder - decoders.0.channel_proj": 12168,
72
+ "Decoder - decoders.0.convnext_block.dwconv": 3600,
73
+ "Decoder - decoders.0.convnext_block.norm": 144,
74
+ "Decoder - decoders.0.convnext_block.pwconv1": 21024,
75
+ "Decoder - decoders.0.convnext_block.pwconv2": 20808,
76
+ "Decoder - decoders.1.channel_proj": 5808,
77
+ "Decoder - decoders.1.convnext_block.dwconv": 2400,
78
+ "Decoder - decoders.1.convnext_block.norm": 96,
79
+ "Decoder - decoders.1.convnext_block.pwconv1": 9408,
80
+ "Decoder - decoders.1.convnext_block.pwconv2": 9264,
81
+ "Decoder - decoders.2.channel_proj": 1752,
82
+ "Decoder - decoders.2.convnext_block.dwconv": 1200,
83
+ "Decoder - decoders.2.convnext_block.norm": 48,
84
+ "Decoder - decoders.2.convnext_block.pwconv1": 2400,
85
+ "Decoder - decoders.2.convnext_block.pwconv2": 2328,
86
+ "Other - final_conv": 75
87
+ },
88
+ "latent_dimensions": {
89
+ "Level_0": {
90
+ "channels": 24,
91
+ "height": 96,
92
+ "width": 96,
93
+ "total_elements": 221184
94
+ },
95
+ "Level_1": {
96
+ "channels": 48,
97
+ "height": 48,
98
+ "width": 48,
99
+ "total_elements": 110592
100
+ },
101
+ "Level_2": {
102
+ "channels": 72,
103
+ "height": 24,
104
+ "width": 24,
105
+ "total_elements": 41472
106
+ },
107
+ "Level_3": {
108
+ "channels": 96,
109
+ "height": 12,
110
+ "width": 12,
111
+ "total_elements": 13824
112
+ }
113
+ },
114
+ "skip_dimensions": {
115
+ "Skip_0": {
116
+ "channels": 24,
117
+ "height": 96,
118
+ "width": 96,
119
+ "total_elements": 221184
120
+ },
121
+ "Skip_1": {
122
+ "channels": 48,
123
+ "height": 48,
124
+ "width": 48,
125
+ "total_elements": 110592
126
+ },
127
+ "Skip_2": {
128
+ "channels": 72,
129
+ "height": 24,
130
+ "width": 24,
131
+ "total_elements": 41472
132
+ }
133
+ },
134
+ "memory_analysis": {
135
+ "peak_memory_inference_mb": 1.7578125,
136
+ "peak_memory_training_mb": 3.515625,
137
+ "peak_elements": 460800
138
+ }
139
+ },
140
+ "timestamp": "2025-11-06T18:07:38.908492"
141
+ }
exp_1762448858_8082_s96_f24_d3_m1_2_3_4/pipeline_results.json ADDED
@@ -0,0 +1,141 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "experiment_id": "exp_1762448858_7941_s96_f24_d3_m1_2_3_4",
3
+ "config": {
4
+ "input_size": 96,
5
+ "base_filters": 24,
6
+ "depth": 3,
7
+ "channel_multipliers": [
8
+ 1,
9
+ 2,
10
+ 3,
11
+ 4
12
+ ],
13
+ "n_channels": 8,
14
+ "n_classes": 3
15
+ },
16
+ "start_time": "2025-11-06T18:07:38.900499",
17
+ "success": true,
18
+ "timings": {
19
+ "onnx_conversion": 0.2930171131156385,
20
+ "openvino_conversion": 4.163721205899492,
21
+ "inference_total": 2.705671763047576,
22
+ "inference_actual": 0.123546,
23
+ "total": 7.187887142878026
24
+ },
25
+ "errors": [],
26
+ "architecture_stats": {
27
+ "input_dimension": 73728,
28
+ "output_dimension": 27648,
29
+ "parameter_distribution": {
30
+ "encoder_params": 77664,
31
+ "decoder_params": 159480,
32
+ "bottleneck_params": 86208,
33
+ "other_params": 75,
34
+ "total_params": 323427,
35
+ "encoder_percentage": 24.01283751820349,
36
+ "decoder_percentage": 49.30942685675593,
37
+ "bottleneck_percentage": 26.654546466436013,
38
+ "other_percentage": 0.02318915860456919
39
+ },
40
+ "component_breakdown": {
41
+ "Encoder - encoders.0.channel_proj": 216,
42
+ "Encoder - encoders.0.convnext_block.dwconv": 1200,
43
+ "Encoder - encoders.0.convnext_block.norm": 48,
44
+ "Encoder - encoders.0.convnext_block.pwconv1": 2400,
45
+ "Encoder - encoders.0.convnext_block.pwconv2": 2328,
46
+ "Encoder - encoders.1.channel_proj": 1200,
47
+ "Encoder - encoders.1.convnext_block.dwconv": 2400,
48
+ "Encoder - encoders.1.convnext_block.norm": 96,
49
+ "Encoder - encoders.1.convnext_block.pwconv1": 9408,
50
+ "Encoder - encoders.1.convnext_block.pwconv2": 9264,
51
+ "Encoder - encoders.2.channel_proj": 3528,
52
+ "Encoder - encoders.2.convnext_block.dwconv": 3600,
53
+ "Encoder - encoders.2.convnext_block.norm": 144,
54
+ "Encoder - encoders.2.convnext_block.pwconv1": 21024,
55
+ "Encoder - encoders.2.convnext_block.pwconv2": 20808,
56
+ "Bottleneck - bottleneck.channel_proj": 7008,
57
+ "Bottleneck - bottleneck.convnext_block.dwconv": 4800,
58
+ "Bottleneck - bottleneck.convnext_block.norm": 192,
59
+ "Bottleneck - bottleneck.convnext_block.pwconv1": 37248,
60
+ "Bottleneck - bottleneck.convnext_block.pwconv2": 36960,
61
+ "Decoder - upsamplers.0": 36960,
62
+ "Decoder - upsamplers.1": 20808,
63
+ "Decoder - upsamplers.2": 9264,
64
+ "Decoder - decoders.0.channel_proj": 12168,
65
+ "Decoder - decoders.0.convnext_block.dwconv": 3600,
66
+ "Decoder - decoders.0.convnext_block.norm": 144,
67
+ "Decoder - decoders.0.convnext_block.pwconv1": 21024,
68
+ "Decoder - decoders.0.convnext_block.pwconv2": 20808,
69
+ "Decoder - decoders.1.channel_proj": 5808,
70
+ "Decoder - decoders.1.convnext_block.dwconv": 2400,
71
+ "Decoder - decoders.1.convnext_block.norm": 96,
72
+ "Decoder - decoders.1.convnext_block.pwconv1": 9408,
73
+ "Decoder - decoders.1.convnext_block.pwconv2": 9264,
74
+ "Decoder - decoders.2.channel_proj": 1752,
75
+ "Decoder - decoders.2.convnext_block.dwconv": 1200,
76
+ "Decoder - decoders.2.convnext_block.norm": 48,
77
+ "Decoder - decoders.2.convnext_block.pwconv1": 2400,
78
+ "Decoder - decoders.2.convnext_block.pwconv2": 2328,
79
+ "Other - final_conv": 75
80
+ },
81
+ "latent_dimensions": {
82
+ "Level_0": {
83
+ "channels": 24,
84
+ "height": 96,
85
+ "width": 96,
86
+ "total_elements": 221184
87
+ },
88
+ "Level_1": {
89
+ "channels": 48,
90
+ "height": 48,
91
+ "width": 48,
92
+ "total_elements": 110592
93
+ },
94
+ "Level_2": {
95
+ "channels": 72,
96
+ "height": 24,
97
+ "width": 24,
98
+ "total_elements": 41472
99
+ },
100
+ "Level_3": {
101
+ "channels": 96,
102
+ "height": 12,
103
+ "width": 12,
104
+ "total_elements": 13824
105
+ }
106
+ },
107
+ "skip_dimensions": {
108
+ "Skip_0": {
109
+ "channels": 24,
110
+ "height": 96,
111
+ "width": 96,
112
+ "total_elements": 221184
113
+ },
114
+ "Skip_1": {
115
+ "channels": 48,
116
+ "height": 48,
117
+ "width": 48,
118
+ "total_elements": 110592
119
+ },
120
+ "Skip_2": {
121
+ "channels": 72,
122
+ "height": 24,
123
+ "width": 24,
124
+ "total_elements": 41472
125
+ }
126
+ },
127
+ "memory_analysis": {
128
+ "peak_memory_inference_mb": 1.7578125,
129
+ "peak_memory_training_mb": 3.515625,
130
+ "peak_elements": 460800
131
+ }
132
+ },
133
+ "inference_results": {
134
+ "success": true,
135
+ "total_time": 2.705671763047576,
136
+ "inference_time": 0.123546,
137
+ "stdout": "[setupvars.sh] OpenVINO environment initialized\nStarting Myriad inference...\nInput: /home/mount/experiments/exp_1762448858_8082_s96_f24_d3_m1_2_3_4/onnx/sample_input.npy\nModel XML: /home/mount/experiments/exp_1762448858_8082_s96_f24_d3_m1_2_3_4/openvino/model.xml\nModel BIN: /home/mount/experiments/exp_1762448858_8082_s96_f24_d3_m1_2_3_4/openvino/model.bin\n[OK] OpenVINO inference engine imported successfully\nDevice: MYRIAD\nModel XML: /home/mount/experiments/exp_1762448858_8082_s96_f24_d3_m1_2_3_4/openvino/model.xml\nModel BIN: /home/mount/experiments/exp_1762448858_8082_s96_f24_d3_m1_2_3_4/openvino/model.bin\nInput file: /home/mount/experiments/exp_1762448858_8082_s96_f24_d3_m1_2_3_4/onnx/sample_input.npy\nInitializing OpenVINO Runtime Core...\nAvailable devices:\n[E:] [BSL] found 0 ioexpander device\n ['CPU', 'GNA', 'MYRIAD']\nLoading network...\nInput blob: input\nInput shape: [1, 8, 96, 96]\nOutput blob: Conv_107\nOutput shape: [1, 3, 96, 96]\nLoading network to MYRIAD...\nLoading input data...\nInput data shape: (1, 8, 96, 96)\nInput data type: float32\nRunning inference...\n[OK] Inference completed!\nInference time: 0.123546 seconds\nOutput shape: (1, 3, 96, 96)\nOutput dtype: float32\nOutput range: [-0.454834, 0.553711]\nMyriad inference completed!\n",
138
+ "stderr": ""
139
+ },
140
+ "end_time": "2025-11-06T18:07:46.088372"
141
+ }
exp_1762448858_8082_s96_f24_d3_m1_2_3_4/run_inference.sh ADDED
@@ -0,0 +1,19 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #!/bin/bash
2
+ # Source OpenVINO environment
3
+ source /opt/intel/openvino/bin/setupvars.sh
4
+
5
+ PYPATH=/home/mount/scripts/simple_inference.py
6
+
7
+ echo "Starting Myriad inference..."
8
+ echo "Input: /home/mount/experiments/exp_1762448858_8082_s96_f24_d3_m1_2_3_4/onnx/sample_input.npy"
9
+ echo "Model XML: /home/mount/experiments/exp_1762448858_8082_s96_f24_d3_m1_2_3_4/openvino/model.xml"
10
+ echo "Model BIN: /home/mount/experiments/exp_1762448858_8082_s96_f24_d3_m1_2_3_4/openvino/model.bin"
11
+
12
+ python3 $PYPATH \
13
+ --input_filepath /home/mount/experiments/exp_1762448858_8082_s96_f24_d3_m1_2_3_4/onnx/sample_input.npy \
14
+ --device_name MYRIAD \
15
+ --model_bin /home/mount/experiments/exp_1762448858_8082_s96_f24_d3_m1_2_3_4/openvino/model.bin \
16
+ --model_xml /home/mount/experiments/exp_1762448858_8082_s96_f24_d3_m1_2_3_4/openvino/model.xml \
17
+ --output /home/mount/experiments/exp_1762448858_8082_s96_f24_d3_m1_2_3_4/inference/
18
+
19
+ echo "Myriad inference completed!"
exp_1762448908_9161_s96_f26_d3_m2_2_2_3/convert_openvino.sh ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #!/bin/bash
2
+ # Source OpenVINO environment
3
+ source /opt/intel/openvino/bin/setupvars.sh
4
+
5
+ # OpenVINO Model Optimizer script for ONNX model
6
+ PYPATH=/opt/intel/openvino_2020.3.194/deployment_tools/model_optimizer/mo.py
7
+ ONNX_MODEL=/home/mount/experiments/exp_1762448908_9161_s96_f26_d3_m2_2_2_3/onnx/model.onnx
8
+ OUTPUT_DIR=/home/mount/experiments/exp_1762448908_9161_s96_f26_d3_m2_2_2_3/openvino
9
+
10
+ echo "Converting ONNX model to OpenVINO IR format..."
11
+ echo "Input model: $ONNX_MODEL"
12
+ echo "Output directory: $OUTPUT_DIR"
13
+
14
+ # Create output directory if it doesn't exist
15
+ mkdir -p "$OUTPUT_DIR"
16
+
17
+ python3 $PYPATH \
18
+ --input_model "$ONNX_MODEL" \
19
+ --data_type FP16 \
20
+ --input_shape "[1,8,96,96]" \
21
+ --mean_values "[0,0,0,0,0,0,0,0]" \
22
+ --scale_values "[1,1,1,1,1,1,1,1]" \
23
+ --progress \
24
+ --stream_output \
25
+ --output_dir "$OUTPUT_DIR" \
26
+ --model_name model
27
+
28
+ echo "OpenVINO conversion completed!"
29
+ echo "Generated files in $OUTPUT_DIR:"
30
+ ls -la "$OUTPUT_DIR/"
exp_1762448908_9161_s96_f26_d3_m2_2_2_3/experiment.log ADDED
@@ -0,0 +1,99 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2025-11-06 18:08:28 - INFO - === Model Creation Phase ===
2
+ 2025-11-06 18:08:28 - INFO - Creating UNet model with config: {'input_size': 96, 'base_filters': 26, 'depth': 3, 'channel_multipliers': [2, 2, 2, 3], 'n_channels': 8, 'n_classes': 3}
3
+ 2025-11-06 18:08:28 - INFO - Model created successfully:
4
+ 2025-11-06 18:08:28 - INFO - Depth: 3
5
+ 2025-11-06 18:08:28 - INFO - Channels per level: [52, 52, 52, 78]
6
+ 2025-11-06 18:08:28 - INFO - Total parameters: 269,727
7
+ 2025-11-06 18:08:28 - INFO - Parameter memory: 1.03 MB
8
+ 2025-11-06 18:08:28 - INFO - === Detailed Architecture Analysis ===
9
+ 2025-11-06 18:08:28 - INFO - Input vector dimension: 73,728 (8 × 96 × 96)
10
+ 2025-11-06 18:08:28 - INFO - Component-wise parameter breakdown:
11
+ 2025-11-06 18:08:28 - INFO - Encoder - encoders.0.channel_proj: 468 parameters
12
+ 2025-11-06 18:08:28 - INFO - Encoder - encoders.0.convnext_block.dwconv: 2,600 parameters
13
+ 2025-11-06 18:08:28 - INFO - Encoder - encoders.0.convnext_block.norm: 104 parameters
14
+ 2025-11-06 18:08:28 - INFO - Encoder - encoders.0.convnext_block.pwconv1: 11,024 parameters
15
+ 2025-11-06 18:08:28 - INFO - Encoder - encoders.0.convnext_block.pwconv2: 10,868 parameters
16
+ 2025-11-06 18:08:28 - INFO - Encoder - encoders.1.convnext_block.dwconv: 2,600 parameters
17
+ 2025-11-06 18:08:28 - INFO - Encoder - encoders.1.convnext_block.norm: 104 parameters
18
+ 2025-11-06 18:08:28 - INFO - Encoder - encoders.1.convnext_block.pwconv1: 11,024 parameters
19
+ 2025-11-06 18:08:28 - INFO - Encoder - encoders.1.convnext_block.pwconv2: 10,868 parameters
20
+ 2025-11-06 18:08:28 - INFO - Encoder - encoders.2.convnext_block.dwconv: 2,600 parameters
21
+ 2025-11-06 18:08:28 - INFO - Encoder - encoders.2.convnext_block.norm: 104 parameters
22
+ 2025-11-06 18:08:28 - INFO - Encoder - encoders.2.convnext_block.pwconv1: 11,024 parameters
23
+ 2025-11-06 18:08:28 - INFO - Encoder - encoders.2.convnext_block.pwconv2: 10,868 parameters
24
+ 2025-11-06 18:08:28 - INFO - Bottleneck - bottleneck.channel_proj: 4,134 parameters
25
+ 2025-11-06 18:08:28 - INFO - Bottleneck - bottleneck.convnext_block.dwconv: 3,900 parameters
26
+ 2025-11-06 18:08:28 - INFO - Bottleneck - bottleneck.convnext_block.norm: 156 parameters
27
+ 2025-11-06 18:08:28 - INFO - Bottleneck - bottleneck.convnext_block.pwconv1: 24,648 parameters
28
+ 2025-11-06 18:08:28 - INFO - Bottleneck - bottleneck.convnext_block.pwconv2: 24,414 parameters
29
+ 2025-11-06 18:08:28 - INFO - Decoder - upsamplers.0: 24,414 parameters
30
+ 2025-11-06 18:08:28 - INFO - Decoder - upsamplers.1: 10,868 parameters
31
+ 2025-11-06 18:08:28 - INFO - Decoder - upsamplers.2: 10,868 parameters
32
+ 2025-11-06 18:08:28 - INFO - Decoder - decoders.0.channel_proj: 6,812 parameters
33
+ 2025-11-06 18:08:28 - INFO - Decoder - decoders.0.convnext_block.dwconv: 2,600 parameters
34
+ 2025-11-06 18:08:28 - INFO - Decoder - decoders.0.convnext_block.norm: 104 parameters
35
+ 2025-11-06 18:08:28 - INFO - Decoder - decoders.0.convnext_block.pwconv1: 11,024 parameters
36
+ 2025-11-06 18:08:28 - INFO - Decoder - decoders.0.convnext_block.pwconv2: 10,868 parameters
37
+ 2025-11-06 18:08:28 - INFO - Decoder - decoders.1.channel_proj: 5,460 parameters
38
+ 2025-11-06 18:08:28 - INFO - Decoder - decoders.1.convnext_block.dwconv: 2,600 parameters
39
+ 2025-11-06 18:08:28 - INFO - Decoder - decoders.1.convnext_block.norm: 104 parameters
40
+ 2025-11-06 18:08:28 - INFO - Decoder - decoders.1.convnext_block.pwconv1: 11,024 parameters
41
+ 2025-11-06 18:08:28 - INFO - Decoder - decoders.1.convnext_block.pwconv2: 10,868 parameters
42
+ 2025-11-06 18:08:28 - INFO - Decoder - decoders.2.channel_proj: 5,460 parameters
43
+ 2025-11-06 18:08:28 - INFO - Decoder - decoders.2.convnext_block.dwconv: 2,600 parameters
44
+ 2025-11-06 18:08:28 - INFO - Decoder - decoders.2.convnext_block.norm: 104 parameters
45
+ 2025-11-06 18:08:28 - INFO - Decoder - decoders.2.convnext_block.pwconv1: 11,024 parameters
46
+ 2025-11-06 18:08:28 - INFO - Decoder - decoders.2.convnext_block.pwconv2: 10,868 parameters
47
+ 2025-11-06 18:08:28 - INFO - Other - final_conv: 159 parameters
48
+ 2025-11-06 18:08:28 - INFO - Parameter distribution summary:
49
+ 2025-11-06 18:08:28 - INFO - Encoder parameters: 74,256 (27.6%)
50
+ 2025-11-06 18:08:28 - INFO - Decoder parameters: 137,670 (51.1%)
51
+ 2025-11-06 18:08:28 - INFO - Bottleneck parameters: 57,252 (21.3%)
52
+ 2025-11-06 18:08:28 - INFO - Other parameters: 159 (0.1%)
53
+ 2025-11-06 18:08:28 - INFO - Latent space dimensions (feature maps at each level):
54
+ 2025-11-06 18:08:28 - INFO - Level 0: 52 × 96 × 96 = 479,232 elements
55
+ 2025-11-06 18:08:28 - INFO - Level 1: 52 × 48 × 48 = 119,808 elements
56
+ 2025-11-06 18:08:28 - INFO - Level 2: 52 × 24 × 24 = 29,952 elements
57
+ 2025-11-06 18:08:28 - INFO - Level 3: 78 × 12 × 12 = 11,232 elements
58
+ 2025-11-06 18:08:28 - INFO - Skip connection dimensions:
59
+ 2025-11-06 18:08:28 - INFO - Skip 0: 52 × 96 × 96 = 479,232 elements
60
+ 2025-11-06 18:08:28 - INFO - Skip 1: 52 × 48 × 48 = 119,808 elements
61
+ 2025-11-06 18:08:28 - INFO - Skip 2: 52 × 24 × 24 = 29,952 elements
62
+ 2025-11-06 18:08:28 - INFO - Memory analysis:
63
+ 2025-11-06 18:08:28 - INFO - Peak feature map memory (inference): 2.72 MB
64
+ 2025-11-06 18:08:28 - INFO - Peak feature map memory (training): 5.45 MB (with gradients)
65
+ 2025-11-06 18:08:28 - INFO - Output vector dimension: 27,648 (3 × 96 × 96)
66
+ 2025-11-06 18:08:28 - INFO - PyTorch model saved to: /home/philab/Desktop/hydranet/experiments/exp_1762448908_9161_s96_f26_d3_m2_2_2_3/pytorch/model.pt
67
+ 2025-11-06 18:08:28 - INFO - === ONNX Conversion Phase ===
68
+ 2025-11-06 18:08:28 - INFO - === Model Export Diagnostics ===
69
+ 2025-11-06 18:08:28 - INFO - PyTorch version: 1.9.0+cu102
70
+ 2025-11-06 18:08:28 - INFO - Model parameters: 269,727
71
+ 2025-11-06 18:08:28 - INFO - Model memory: 1.03 MB
72
+ 2025-11-06 18:08:28 - INFO - Starting ONNX export with opset version 11
73
+ 2025-11-06 18:08:28 - INFO - Model input shape: torch.Size([1, 8, 96, 96])
74
+ 2025-11-06 18:08:28 - INFO - Model input dtype: torch.float32
75
+ 2025-11-06 18:08:28 - INFO - Forward pass successful. Output shape: torch.Size([1, 3, 96, 96])
76
+ 2025-11-06 18:08:28 - INFO - Output dtype: torch.float32
77
+ 2025-11-06 18:08:28 - INFO - Output value range: [-0.5734, 0.5689]
78
+ 2025-11-06 18:08:29 - INFO - Model successfully exported to /home/philab/Desktop/hydranet/experiments/exp_1762448908_9161_s96_f26_d3_m2_2_2_3/onnx/model.onnx
79
+ 2025-11-06 18:08:29 - INFO - ONNX model size: 1.04 MB
80
+ 2025-11-06 18:08:29 - INFO - Saved dummy input with shape (1, 8, 96, 96) to /home/philab/Desktop/hydranet/experiments/exp_1762448908_9161_s96_f26_d3_m2_2_2_3/onnx/sample_input.npy
81
+ 2025-11-06 18:08:29 - INFO - Input data type: float32
82
+ 2025-11-06 18:08:29 - INFO - Input value range: [-3.9566, 4.1282]
83
+ 2025-11-06 18:08:29 - INFO - === OpenVINO Conversion Phase ===
84
+ 2025-11-06 18:08:29 - INFO - Starting OpenVINO conversion in Docker container...
85
+ 2025-11-06 18:08:33 - INFO - OpenVINO conversion completed in 3.98 seconds
86
+ 2025-11-06 18:08:33 - INFO - OpenVINO model files created:
87
+ 2025-11-06 18:08:33 - INFO - XML file: /home/philab/Desktop/hydranet/experiments/exp_1762448908_9161_s96_f26_d3_m2_2_2_3/openvino/model.xml (0.08 MB)
88
+ 2025-11-06 18:08:33 - INFO - BIN file: /home/philab/Desktop/hydranet/experiments/exp_1762448908_9161_s96_f26_d3_m2_2_2_3/openvino/model.bin (0.51 MB)
89
+ 2025-11-06 18:08:33 - INFO - === Myriad Inference Phase ===
90
+ 2025-11-06 18:08:33 - INFO - Starting Myriad inference in Docker container...
91
+ 2025-11-06 18:08:36 - INFO - Myriad inference completed in 3.00 seconds
92
+ 2025-11-06 18:08:36 - INFO - Actual inference time: 0.351092 seconds
93
+ 2025-11-06 18:08:36 - INFO - ✅ Complete pipeline executed successfully!
94
+ 2025-11-06 18:08:36 - INFO - ✅ Experiment 60 completed successfully
95
+ 2025-11-06 18:08:36 - INFO - Inference time: 0.351092s
96
+ 2025-11-06 18:08:36 - INFO -
97
+ === Experiment 61/2475 ===
98
+ 2025-11-06 18:08:36 - INFO - Experiment ID: exp_1762448916_9432_s96_f26_d3_m1_2_3_3
99
+ 2025-11-06 18:08:36 - INFO - Experiment directory: /home/philab/Desktop/hydranet/experiments/exp_1762448916_9432_s96_f26_d3_m1_2_3_3
exp_1762448908_9161_s96_f26_d3_m2_2_2_3/model_info.json ADDED
@@ -0,0 +1,139 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "config": {
3
+ "input_size": 96,
4
+ "base_filters": 26,
5
+ "depth": 3,
6
+ "channel_multipliers": [
7
+ 2,
8
+ 2,
9
+ 2,
10
+ 3
11
+ ],
12
+ "n_channels": 8,
13
+ "n_classes": 3
14
+ },
15
+ "model_info": {
16
+ "depth": 3,
17
+ "channels_per_level": [
18
+ 52,
19
+ 52,
20
+ 52,
21
+ 78
22
+ ],
23
+ "channel_multipliers": [
24
+ 2,
25
+ 2,
26
+ 2,
27
+ 3
28
+ ],
29
+ "total_parameters": 269727,
30
+ "trainable_parameters": 269727,
31
+ "model_size_mb": 1.0289268493652344
32
+ },
33
+ "architecture_stats": {
34
+ "input_dimension": 73728,
35
+ "output_dimension": 27648,
36
+ "parameter_distribution": {
37
+ "encoder_params": 74256,
38
+ "decoder_params": 137670,
39
+ "bottleneck_params": 57252,
40
+ "other_params": 159,
41
+ "total_params": 269337,
42
+ "encoder_percentage": 27.569921696610567,
43
+ "decoder_percentage": 51.11440314550173,
44
+ "bottleneck_percentage": 21.256641308101003,
45
+ "other_percentage": 0.05903384978669844
46
+ },
47
+ "component_breakdown": {
48
+ "Encoder - encoders.0.channel_proj": 468,
49
+ "Encoder - encoders.0.convnext_block.dwconv": 2600,
50
+ "Encoder - encoders.0.convnext_block.norm": 104,
51
+ "Encoder - encoders.0.convnext_block.pwconv1": 11024,
52
+ "Encoder - encoders.0.convnext_block.pwconv2": 10868,
53
+ "Encoder - encoders.1.convnext_block.dwconv": 2600,
54
+ "Encoder - encoders.1.convnext_block.norm": 104,
55
+ "Encoder - encoders.1.convnext_block.pwconv1": 11024,
56
+ "Encoder - encoders.1.convnext_block.pwconv2": 10868,
57
+ "Encoder - encoders.2.convnext_block.dwconv": 2600,
58
+ "Encoder - encoders.2.convnext_block.norm": 104,
59
+ "Encoder - encoders.2.convnext_block.pwconv1": 11024,
60
+ "Encoder - encoders.2.convnext_block.pwconv2": 10868,
61
+ "Bottleneck - bottleneck.channel_proj": 4134,
62
+ "Bottleneck - bottleneck.convnext_block.dwconv": 3900,
63
+ "Bottleneck - bottleneck.convnext_block.norm": 156,
64
+ "Bottleneck - bottleneck.convnext_block.pwconv1": 24648,
65
+ "Bottleneck - bottleneck.convnext_block.pwconv2": 24414,
66
+ "Decoder - upsamplers.0": 24414,
67
+ "Decoder - upsamplers.1": 10868,
68
+ "Decoder - upsamplers.2": 10868,
69
+ "Decoder - decoders.0.channel_proj": 6812,
70
+ "Decoder - decoders.0.convnext_block.dwconv": 2600,
71
+ "Decoder - decoders.0.convnext_block.norm": 104,
72
+ "Decoder - decoders.0.convnext_block.pwconv1": 11024,
73
+ "Decoder - decoders.0.convnext_block.pwconv2": 10868,
74
+ "Decoder - decoders.1.channel_proj": 5460,
75
+ "Decoder - decoders.1.convnext_block.dwconv": 2600,
76
+ "Decoder - decoders.1.convnext_block.norm": 104,
77
+ "Decoder - decoders.1.convnext_block.pwconv1": 11024,
78
+ "Decoder - decoders.1.convnext_block.pwconv2": 10868,
79
+ "Decoder - decoders.2.channel_proj": 5460,
80
+ "Decoder - decoders.2.convnext_block.dwconv": 2600,
81
+ "Decoder - decoders.2.convnext_block.norm": 104,
82
+ "Decoder - decoders.2.convnext_block.pwconv1": 11024,
83
+ "Decoder - decoders.2.convnext_block.pwconv2": 10868,
84
+ "Other - final_conv": 159
85
+ },
86
+ "latent_dimensions": {
87
+ "Level_0": {
88
+ "channels": 52,
89
+ "height": 96,
90
+ "width": 96,
91
+ "total_elements": 479232
92
+ },
93
+ "Level_1": {
94
+ "channels": 52,
95
+ "height": 48,
96
+ "width": 48,
97
+ "total_elements": 119808
98
+ },
99
+ "Level_2": {
100
+ "channels": 52,
101
+ "height": 24,
102
+ "width": 24,
103
+ "total_elements": 29952
104
+ },
105
+ "Level_3": {
106
+ "channels": 78,
107
+ "height": 12,
108
+ "width": 12,
109
+ "total_elements": 11232
110
+ }
111
+ },
112
+ "skip_dimensions": {
113
+ "Skip_0": {
114
+ "channels": 52,
115
+ "height": 96,
116
+ "width": 96,
117
+ "total_elements": 479232
118
+ },
119
+ "Skip_1": {
120
+ "channels": 52,
121
+ "height": 48,
122
+ "width": 48,
123
+ "total_elements": 119808
124
+ },
125
+ "Skip_2": {
126
+ "channels": 52,
127
+ "height": 24,
128
+ "width": 24,
129
+ "total_elements": 29952
130
+ }
131
+ },
132
+ "memory_analysis": {
133
+ "peak_memory_inference_mb": 2.7235107421875,
134
+ "peak_memory_training_mb": 5.447021484375,
135
+ "peak_elements": 713952
136
+ }
137
+ },
138
+ "timestamp": "2025-11-06T18:08:28.837227"
139
+ }
exp_1762448908_9161_s96_f26_d3_m2_2_2_3/pipeline_results.json ADDED
@@ -0,0 +1,139 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "experiment_id": "exp_1762448908_7968_s96_f26_d3_m2_2_2_3",
3
+ "config": {
4
+ "input_size": 96,
5
+ "base_filters": 26,
6
+ "depth": 3,
7
+ "channel_multipliers": [
8
+ 2,
9
+ 2,
10
+ 2,
11
+ 3
12
+ ],
13
+ "n_channels": 8,
14
+ "n_classes": 3
15
+ },
16
+ "start_time": "2025-11-06T18:08:28.830729",
17
+ "success": true,
18
+ "timings": {
19
+ "onnx_conversion": 0.29142461507581174,
20
+ "openvino_conversion": 3.984111813828349,
21
+ "inference_total": 3.001626858022064,
22
+ "inference_actual": 0.351092,
23
+ "total": 7.291057776892558
24
+ },
25
+ "errors": [],
26
+ "architecture_stats": {
27
+ "input_dimension": 73728,
28
+ "output_dimension": 27648,
29
+ "parameter_distribution": {
30
+ "encoder_params": 74256,
31
+ "decoder_params": 137670,
32
+ "bottleneck_params": 57252,
33
+ "other_params": 159,
34
+ "total_params": 269337,
35
+ "encoder_percentage": 27.569921696610567,
36
+ "decoder_percentage": 51.11440314550173,
37
+ "bottleneck_percentage": 21.256641308101003,
38
+ "other_percentage": 0.05903384978669844
39
+ },
40
+ "component_breakdown": {
41
+ "Encoder - encoders.0.channel_proj": 468,
42
+ "Encoder - encoders.0.convnext_block.dwconv": 2600,
43
+ "Encoder - encoders.0.convnext_block.norm": 104,
44
+ "Encoder - encoders.0.convnext_block.pwconv1": 11024,
45
+ "Encoder - encoders.0.convnext_block.pwconv2": 10868,
46
+ "Encoder - encoders.1.convnext_block.dwconv": 2600,
47
+ "Encoder - encoders.1.convnext_block.norm": 104,
48
+ "Encoder - encoders.1.convnext_block.pwconv1": 11024,
49
+ "Encoder - encoders.1.convnext_block.pwconv2": 10868,
50
+ "Encoder - encoders.2.convnext_block.dwconv": 2600,
51
+ "Encoder - encoders.2.convnext_block.norm": 104,
52
+ "Encoder - encoders.2.convnext_block.pwconv1": 11024,
53
+ "Encoder - encoders.2.convnext_block.pwconv2": 10868,
54
+ "Bottleneck - bottleneck.channel_proj": 4134,
55
+ "Bottleneck - bottleneck.convnext_block.dwconv": 3900,
56
+ "Bottleneck - bottleneck.convnext_block.norm": 156,
57
+ "Bottleneck - bottleneck.convnext_block.pwconv1": 24648,
58
+ "Bottleneck - bottleneck.convnext_block.pwconv2": 24414,
59
+ "Decoder - upsamplers.0": 24414,
60
+ "Decoder - upsamplers.1": 10868,
61
+ "Decoder - upsamplers.2": 10868,
62
+ "Decoder - decoders.0.channel_proj": 6812,
63
+ "Decoder - decoders.0.convnext_block.dwconv": 2600,
64
+ "Decoder - decoders.0.convnext_block.norm": 104,
65
+ "Decoder - decoders.0.convnext_block.pwconv1": 11024,
66
+ "Decoder - decoders.0.convnext_block.pwconv2": 10868,
67
+ "Decoder - decoders.1.channel_proj": 5460,
68
+ "Decoder - decoders.1.convnext_block.dwconv": 2600,
69
+ "Decoder - decoders.1.convnext_block.norm": 104,
70
+ "Decoder - decoders.1.convnext_block.pwconv1": 11024,
71
+ "Decoder - decoders.1.convnext_block.pwconv2": 10868,
72
+ "Decoder - decoders.2.channel_proj": 5460,
73
+ "Decoder - decoders.2.convnext_block.dwconv": 2600,
74
+ "Decoder - decoders.2.convnext_block.norm": 104,
75
+ "Decoder - decoders.2.convnext_block.pwconv1": 11024,
76
+ "Decoder - decoders.2.convnext_block.pwconv2": 10868,
77
+ "Other - final_conv": 159
78
+ },
79
+ "latent_dimensions": {
80
+ "Level_0": {
81
+ "channels": 52,
82
+ "height": 96,
83
+ "width": 96,
84
+ "total_elements": 479232
85
+ },
86
+ "Level_1": {
87
+ "channels": 52,
88
+ "height": 48,
89
+ "width": 48,
90
+ "total_elements": 119808
91
+ },
92
+ "Level_2": {
93
+ "channels": 52,
94
+ "height": 24,
95
+ "width": 24,
96
+ "total_elements": 29952
97
+ },
98
+ "Level_3": {
99
+ "channels": 78,
100
+ "height": 12,
101
+ "width": 12,
102
+ "total_elements": 11232
103
+ }
104
+ },
105
+ "skip_dimensions": {
106
+ "Skip_0": {
107
+ "channels": 52,
108
+ "height": 96,
109
+ "width": 96,
110
+ "total_elements": 479232
111
+ },
112
+ "Skip_1": {
113
+ "channels": 52,
114
+ "height": 48,
115
+ "width": 48,
116
+ "total_elements": 119808
117
+ },
118
+ "Skip_2": {
119
+ "channels": 52,
120
+ "height": 24,
121
+ "width": 24,
122
+ "total_elements": 29952
123
+ }
124
+ },
125
+ "memory_analysis": {
126
+ "peak_memory_inference_mb": 2.7235107421875,
127
+ "peak_memory_training_mb": 5.447021484375,
128
+ "peak_elements": 713952
129
+ }
130
+ },
131
+ "inference_results": {
132
+ "success": true,
133
+ "total_time": 3.001626858022064,
134
+ "inference_time": 0.351092,
135
+ "stdout": "[setupvars.sh] OpenVINO environment initialized\nStarting Myriad inference...\nInput: /home/mount/experiments/exp_1762448908_9161_s96_f26_d3_m2_2_2_3/onnx/sample_input.npy\nModel XML: /home/mount/experiments/exp_1762448908_9161_s96_f26_d3_m2_2_2_3/openvino/model.xml\nModel BIN: /home/mount/experiments/exp_1762448908_9161_s96_f26_d3_m2_2_2_3/openvino/model.bin\n[OK] OpenVINO inference engine imported successfully\nDevice: MYRIAD\nModel XML: /home/mount/experiments/exp_1762448908_9161_s96_f26_d3_m2_2_2_3/openvino/model.xml\nModel BIN: /home/mount/experiments/exp_1762448908_9161_s96_f26_d3_m2_2_2_3/openvino/model.bin\nInput file: /home/mount/experiments/exp_1762448908_9161_s96_f26_d3_m2_2_2_3/onnx/sample_input.npy\nInitializing OpenVINO Runtime Core...\nAvailable devices:\n[E:] [BSL] found 0 ioexpander device\n ['CPU', 'GNA', 'MYRIAD']\nLoading network...\nInput blob: input\nInput shape: [1, 8, 96, 96]\nOutput blob: Conv_105\nOutput shape: [1, 3, 96, 96]\nLoading network to MYRIAD...\nLoading input data...\nInput data shape: (1, 8, 96, 96)\nInput data type: float32\nRunning inference...\n[OK] Inference completed!\nInference time: 0.351092 seconds\nOutput shape: (1, 3, 96, 96)\nOutput dtype: float32\nOutput range: [-0.573730, 0.568359]\nMyriad inference completed!\n",
136
+ "stderr": ""
137
+ },
138
+ "end_time": "2025-11-06T18:08:36.121777"
139
+ }
exp_1762448908_9161_s96_f26_d3_m2_2_2_3/run_inference.sh ADDED
@@ -0,0 +1,19 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #!/bin/bash
2
+ # Source OpenVINO environment
3
+ source /opt/intel/openvino/bin/setupvars.sh
4
+
5
+ PYPATH=/home/mount/scripts/simple_inference.py
6
+
7
+ echo "Starting Myriad inference..."
8
+ echo "Input: /home/mount/experiments/exp_1762448908_9161_s96_f26_d3_m2_2_2_3/onnx/sample_input.npy"
9
+ echo "Model XML: /home/mount/experiments/exp_1762448908_9161_s96_f26_d3_m2_2_2_3/openvino/model.xml"
10
+ echo "Model BIN: /home/mount/experiments/exp_1762448908_9161_s96_f26_d3_m2_2_2_3/openvino/model.bin"
11
+
12
+ python3 $PYPATH \
13
+ --input_filepath /home/mount/experiments/exp_1762448908_9161_s96_f26_d3_m2_2_2_3/onnx/sample_input.npy \
14
+ --device_name MYRIAD \
15
+ --model_bin /home/mount/experiments/exp_1762448908_9161_s96_f26_d3_m2_2_2_3/openvino/model.bin \
16
+ --model_xml /home/mount/experiments/exp_1762448908_9161_s96_f26_d3_m2_2_2_3/openvino/model.xml \
17
+ --output /home/mount/experiments/exp_1762448908_9161_s96_f26_d3_m2_2_2_3/inference/
18
+
19
+ echo "Myriad inference completed!"
exp_1762449009_9308_s96_f28_d3_m1_2_2_3/convert_openvino.sh ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #!/bin/bash
2
+ # Source OpenVINO environment
3
+ source /opt/intel/openvino/bin/setupvars.sh
4
+
5
+ # OpenVINO Model Optimizer script for ONNX model
6
+ PYPATH=/opt/intel/openvino_2020.3.194/deployment_tools/model_optimizer/mo.py
7
+ ONNX_MODEL=/home/mount/experiments/exp_1762449009_9308_s96_f28_d3_m1_2_2_3/onnx/model.onnx
8
+ OUTPUT_DIR=/home/mount/experiments/exp_1762449009_9308_s96_f28_d3_m1_2_2_3/openvino
9
+
10
+ echo "Converting ONNX model to OpenVINO IR format..."
11
+ echo "Input model: $ONNX_MODEL"
12
+ echo "Output directory: $OUTPUT_DIR"
13
+
14
+ # Create output directory if it doesn't exist
15
+ mkdir -p "$OUTPUT_DIR"
16
+
17
+ python3 $PYPATH \
18
+ --input_model "$ONNX_MODEL" \
19
+ --data_type FP16 \
20
+ --input_shape "[1,8,96,96]" \
21
+ --mean_values "[0,0,0,0,0,0,0,0]" \
22
+ --scale_values "[1,1,1,1,1,1,1,1]" \
23
+ --progress \
24
+ --stream_output \
25
+ --output_dir "$OUTPUT_DIR" \
26
+ --model_name model
27
+
28
+ echo "OpenVINO conversion completed!"
29
+ echo "Generated files in $OUTPUT_DIR:"
30
+ ls -la "$OUTPUT_DIR/"
exp_1762449009_9308_s96_f28_d3_m1_2_2_3/experiment.log ADDED
@@ -0,0 +1,100 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2025-11-06 18:10:09 - INFO - === Model Creation Phase ===
2
+ 2025-11-06 18:10:09 - INFO - Creating UNet model with config: {'input_size': 96, 'base_filters': 28, 'depth': 3, 'channel_multipliers': [1, 2, 2, 3], 'n_channels': 8, 'n_classes': 3}
3
+ 2025-11-06 18:10:09 - INFO - Model created successfully:
4
+ 2025-11-06 18:10:09 - INFO - Depth: 3
5
+ 2025-11-06 18:10:09 - INFO - Channels per level: [28, 56, 56, 84]
6
+ 2025-11-06 18:10:09 - INFO - Total parameters: 267,319
7
+ 2025-11-06 18:10:09 - INFO - Parameter memory: 1.02 MB
8
+ 2025-11-06 18:10:09 - INFO - === Detailed Architecture Analysis ===
9
+ 2025-11-06 18:10:09 - INFO - Input vector dimension: 73,728 (8 × 96 × 96)
10
+ 2025-11-06 18:10:09 - INFO - Component-wise parameter breakdown:
11
+ 2025-11-06 18:10:09 - INFO - Encoder - encoders.0.channel_proj: 252 parameters
12
+ 2025-11-06 18:10:09 - INFO - Encoder - encoders.0.convnext_block.dwconv: 1,400 parameters
13
+ 2025-11-06 18:10:09 - INFO - Encoder - encoders.0.convnext_block.norm: 56 parameters
14
+ 2025-11-06 18:10:09 - INFO - Encoder - encoders.0.convnext_block.pwconv1: 3,248 parameters
15
+ 2025-11-06 18:10:09 - INFO - Encoder - encoders.0.convnext_block.pwconv2: 3,164 parameters
16
+ 2025-11-06 18:10:09 - INFO - Encoder - encoders.1.channel_proj: 1,624 parameters
17
+ 2025-11-06 18:10:09 - INFO - Encoder - encoders.1.convnext_block.dwconv: 2,800 parameters
18
+ 2025-11-06 18:10:09 - INFO - Encoder - encoders.1.convnext_block.norm: 112 parameters
19
+ 2025-11-06 18:10:09 - INFO - Encoder - encoders.1.convnext_block.pwconv1: 12,768 parameters
20
+ 2025-11-06 18:10:09 - INFO - Encoder - encoders.1.convnext_block.pwconv2: 12,600 parameters
21
+ 2025-11-06 18:10:09 - INFO - Encoder - encoders.2.convnext_block.dwconv: 2,800 parameters
22
+ 2025-11-06 18:10:09 - INFO - Encoder - encoders.2.convnext_block.norm: 112 parameters
23
+ 2025-11-06 18:10:09 - INFO - Encoder - encoders.2.convnext_block.pwconv1: 12,768 parameters
24
+ 2025-11-06 18:10:09 - INFO - Encoder - encoders.2.convnext_block.pwconv2: 12,600 parameters
25
+ 2025-11-06 18:10:09 - INFO - Bottleneck - bottleneck.channel_proj: 4,788 parameters
26
+ 2025-11-06 18:10:09 - INFO - Bottleneck - bottleneck.convnext_block.dwconv: 4,200 parameters
27
+ 2025-11-06 18:10:09 - INFO - Bottleneck - bottleneck.convnext_block.norm: 168 parameters
28
+ 2025-11-06 18:10:09 - INFO - Bottleneck - bottleneck.convnext_block.pwconv1: 28,560 parameters
29
+ 2025-11-06 18:10:09 - INFO - Bottleneck - bottleneck.convnext_block.pwconv2: 28,308 parameters
30
+ 2025-11-06 18:10:09 - INFO - Decoder - upsamplers.0: 28,308 parameters
31
+ 2025-11-06 18:10:09 - INFO - Decoder - upsamplers.1: 12,600 parameters
32
+ 2025-11-06 18:10:09 - INFO - Decoder - upsamplers.2: 12,600 parameters
33
+ 2025-11-06 18:10:09 - INFO - Decoder - decoders.0.channel_proj: 7,896 parameters
34
+ 2025-11-06 18:10:09 - INFO - Decoder - decoders.0.convnext_block.dwconv: 2,800 parameters
35
+ 2025-11-06 18:10:09 - INFO - Decoder - decoders.0.convnext_block.norm: 112 parameters
36
+ 2025-11-06 18:10:09 - INFO - Decoder - decoders.0.convnext_block.pwconv1: 12,768 parameters
37
+ 2025-11-06 18:10:09 - INFO - Decoder - decoders.0.convnext_block.pwconv2: 12,600 parameters
38
+ 2025-11-06 18:10:09 - INFO - Decoder - decoders.1.channel_proj: 6,328 parameters
39
+ 2025-11-06 18:10:09 - INFO - Decoder - decoders.1.convnext_block.dwconv: 2,800 parameters
40
+ 2025-11-06 18:10:09 - INFO - Decoder - decoders.1.convnext_block.norm: 112 parameters
41
+ 2025-11-06 18:10:09 - INFO - Decoder - decoders.1.convnext_block.pwconv1: 12,768 parameters
42
+ 2025-11-06 18:10:09 - INFO - Decoder - decoders.1.convnext_block.pwconv2: 12,600 parameters
43
+ 2025-11-06 18:10:09 - INFO - Decoder - decoders.2.channel_proj: 2,380 parameters
44
+ 2025-11-06 18:10:09 - INFO - Decoder - decoders.2.convnext_block.dwconv: 1,400 parameters
45
+ 2025-11-06 18:10:09 - INFO - Decoder - decoders.2.convnext_block.norm: 56 parameters
46
+ 2025-11-06 18:10:09 - INFO - Decoder - decoders.2.convnext_block.pwconv1: 3,248 parameters
47
+ 2025-11-06 18:10:09 - INFO - Decoder - decoders.2.convnext_block.pwconv2: 3,164 parameters
48
+ 2025-11-06 18:10:09 - INFO - Other - final_conv: 87 parameters
49
+ 2025-11-06 18:10:09 - INFO - Parameter distribution summary:
50
+ 2025-11-06 18:10:09 - INFO - Encoder parameters: 66,304 (24.8%)
51
+ 2025-11-06 18:10:09 - INFO - Decoder parameters: 134,540 (50.4%)
52
+ 2025-11-06 18:10:09 - INFO - Bottleneck parameters: 66,024 (24.7%)
53
+ 2025-11-06 18:10:09 - INFO - Other parameters: 87 (0.0%)
54
+ 2025-11-06 18:10:09 - INFO - Latent space dimensions (feature maps at each level):
55
+ 2025-11-06 18:10:09 - INFO - Level 0: 28 × 96 × 96 = 258,048 elements
56
+ 2025-11-06 18:10:09 - INFO - Level 1: 56 × 48 × 48 = 129,024 elements
57
+ 2025-11-06 18:10:09 - INFO - Level 2: 56 × 24 × 24 = 32,256 elements
58
+ 2025-11-06 18:10:09 - INFO - Level 3: 84 × 12 × 12 = 12,096 elements
59
+ 2025-11-06 18:10:09 - INFO - Skip connection dimensions:
60
+ 2025-11-06 18:10:09 - INFO - Skip 0: 28 × 96 × 96 = 258,048 elements
61
+ 2025-11-06 18:10:09 - INFO - Skip 1: 56 × 48 × 48 = 129,024 elements
62
+ 2025-11-06 18:10:09 - INFO - Skip 2: 56 × 24 × 24 = 32,256 elements
63
+ 2025-11-06 18:10:09 - INFO - Memory analysis:
64
+ 2025-11-06 18:10:09 - INFO - Peak feature map memory (inference): 1.93 MB
65
+ 2025-11-06 18:10:09 - INFO - Peak feature map memory (training): 3.85 MB (with gradients)
66
+ 2025-11-06 18:10:09 - INFO - Output vector dimension: 27,648 (3 × 96 × 96)
67
+ 2025-11-06 18:10:09 - INFO - PyTorch model saved to: /home/philab/Desktop/hydranet/experiments/exp_1762449009_9308_s96_f28_d3_m1_2_2_3/pytorch/model.pt
68
+ 2025-11-06 18:10:09 - INFO - === ONNX Conversion Phase ===
69
+ 2025-11-06 18:10:09 - INFO - === Model Export Diagnostics ===
70
+ 2025-11-06 18:10:09 - INFO - PyTorch version: 1.9.0+cu102
71
+ 2025-11-06 18:10:09 - INFO - Model parameters: 267,319
72
+ 2025-11-06 18:10:09 - INFO - Model memory: 1.02 MB
73
+ 2025-11-06 18:10:09 - INFO - Starting ONNX export with opset version 11
74
+ 2025-11-06 18:10:09 - INFO - Model input shape: torch.Size([1, 8, 96, 96])
75
+ 2025-11-06 18:10:09 - INFO - Model input dtype: torch.float32
76
+ 2025-11-06 18:10:09 - INFO - Forward pass successful. Output shape: torch.Size([1, 3, 96, 96])
77
+ 2025-11-06 18:10:09 - INFO - Output dtype: torch.float32
78
+ 2025-11-06 18:10:09 - INFO - Output value range: [-0.6065, 0.7020]
79
+ 2025-11-06 18:10:10 - INFO - Model successfully exported to /home/philab/Desktop/hydranet/experiments/exp_1762449009_9308_s96_f28_d3_m1_2_2_3/onnx/model.onnx
80
+ 2025-11-06 18:10:10 - INFO - ONNX model size: 1.03 MB
81
+ 2025-11-06 18:10:10 - INFO - Saved dummy input with shape (1, 8, 96, 96) to /home/philab/Desktop/hydranet/experiments/exp_1762449009_9308_s96_f28_d3_m1_2_2_3/onnx/sample_input.npy
82
+ 2025-11-06 18:10:10 - INFO - Input data type: float32
83
+ 2025-11-06 18:10:10 - INFO - Input value range: [-3.8743, 4.3175]
84
+ 2025-11-06 18:10:10 - INFO - === OpenVINO Conversion Phase ===
85
+ 2025-11-06 18:10:10 - INFO - Starting OpenVINO conversion in Docker container...
86
+ 2025-11-06 18:10:14 - INFO - OpenVINO conversion completed in 4.10 seconds
87
+ 2025-11-06 18:10:14 - INFO - OpenVINO model files created:
88
+ 2025-11-06 18:10:14 - INFO - XML file: /home/philab/Desktop/hydranet/experiments/exp_1762449009_9308_s96_f28_d3_m1_2_2_3/openvino/model.xml (0.09 MB)
89
+ 2025-11-06 18:10:14 - INFO - BIN file: /home/philab/Desktop/hydranet/experiments/exp_1762449009_9308_s96_f28_d3_m1_2_2_3/openvino/model.bin (0.51 MB)
90
+ 2025-11-06 18:10:14 - INFO - === Myriad Inference Phase ===
91
+ 2025-11-06 18:10:14 - INFO - Starting Myriad inference in Docker container...
92
+ 2025-11-06 18:10:16 - INFO - Myriad inference completed in 2.81 seconds
93
+ 2025-11-06 18:10:16 - INFO - Actual inference time: 0.203952 seconds
94
+ 2025-11-06 18:10:16 - INFO - ✅ Complete pipeline executed successfully!
95
+ 2025-11-06 18:10:16 - INFO - ✅ Experiment 74 completed successfully
96
+ 2025-11-06 18:10:16 - INFO - Inference time: 0.203952s
97
+ 2025-11-06 18:10:16 - INFO -
98
+ === Experiment 75/2475 ===
99
+ 2025-11-06 18:10:16 - INFO - Experiment ID: exp_1762449016_3334_s96_f28_d3_m1_2_3_4
100
+ 2025-11-06 18:10:16 - INFO - Experiment directory: /home/philab/Desktop/hydranet/experiments/exp_1762449016_3334_s96_f28_d3_m1_2_3_4
exp_1762449009_9308_s96_f28_d3_m1_2_2_3/model_info.json ADDED
@@ -0,0 +1,140 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "config": {
3
+ "input_size": 96,
4
+ "base_filters": 28,
5
+ "depth": 3,
6
+ "channel_multipliers": [
7
+ 1,
8
+ 2,
9
+ 2,
10
+ 3
11
+ ],
12
+ "n_channels": 8,
13
+ "n_classes": 3
14
+ },
15
+ "model_info": {
16
+ "depth": 3,
17
+ "channels_per_level": [
18
+ 28,
19
+ 56,
20
+ 56,
21
+ 84
22
+ ],
23
+ "channel_multipliers": [
24
+ 1,
25
+ 2,
26
+ 2,
27
+ 3
28
+ ],
29
+ "total_parameters": 267319,
30
+ "trainable_parameters": 267319,
31
+ "model_size_mb": 1.0197410583496094
32
+ },
33
+ "architecture_stats": {
34
+ "input_dimension": 73728,
35
+ "output_dimension": 27648,
36
+ "parameter_distribution": {
37
+ "encoder_params": 66304,
38
+ "decoder_params": 134540,
39
+ "bottleneck_params": 66024,
40
+ "other_params": 87,
41
+ "total_params": 266955,
42
+ "encoder_percentage": 24.837144837144837,
43
+ "decoder_percentage": 50.39800715476391,
44
+ "bottleneck_percentage": 24.73225824577176,
45
+ "other_percentage": 0.03258976231949205
46
+ },
47
+ "component_breakdown": {
48
+ "Encoder - encoders.0.channel_proj": 252,
49
+ "Encoder - encoders.0.convnext_block.dwconv": 1400,
50
+ "Encoder - encoders.0.convnext_block.norm": 56,
51
+ "Encoder - encoders.0.convnext_block.pwconv1": 3248,
52
+ "Encoder - encoders.0.convnext_block.pwconv2": 3164,
53
+ "Encoder - encoders.1.channel_proj": 1624,
54
+ "Encoder - encoders.1.convnext_block.dwconv": 2800,
55
+ "Encoder - encoders.1.convnext_block.norm": 112,
56
+ "Encoder - encoders.1.convnext_block.pwconv1": 12768,
57
+ "Encoder - encoders.1.convnext_block.pwconv2": 12600,
58
+ "Encoder - encoders.2.convnext_block.dwconv": 2800,
59
+ "Encoder - encoders.2.convnext_block.norm": 112,
60
+ "Encoder - encoders.2.convnext_block.pwconv1": 12768,
61
+ "Encoder - encoders.2.convnext_block.pwconv2": 12600,
62
+ "Bottleneck - bottleneck.channel_proj": 4788,
63
+ "Bottleneck - bottleneck.convnext_block.dwconv": 4200,
64
+ "Bottleneck - bottleneck.convnext_block.norm": 168,
65
+ "Bottleneck - bottleneck.convnext_block.pwconv1": 28560,
66
+ "Bottleneck - bottleneck.convnext_block.pwconv2": 28308,
67
+ "Decoder - upsamplers.0": 28308,
68
+ "Decoder - upsamplers.1": 12600,
69
+ "Decoder - upsamplers.2": 12600,
70
+ "Decoder - decoders.0.channel_proj": 7896,
71
+ "Decoder - decoders.0.convnext_block.dwconv": 2800,
72
+ "Decoder - decoders.0.convnext_block.norm": 112,
73
+ "Decoder - decoders.0.convnext_block.pwconv1": 12768,
74
+ "Decoder - decoders.0.convnext_block.pwconv2": 12600,
75
+ "Decoder - decoders.1.channel_proj": 6328,
76
+ "Decoder - decoders.1.convnext_block.dwconv": 2800,
77
+ "Decoder - decoders.1.convnext_block.norm": 112,
78
+ "Decoder - decoders.1.convnext_block.pwconv1": 12768,
79
+ "Decoder - decoders.1.convnext_block.pwconv2": 12600,
80
+ "Decoder - decoders.2.channel_proj": 2380,
81
+ "Decoder - decoders.2.convnext_block.dwconv": 1400,
82
+ "Decoder - decoders.2.convnext_block.norm": 56,
83
+ "Decoder - decoders.2.convnext_block.pwconv1": 3248,
84
+ "Decoder - decoders.2.convnext_block.pwconv2": 3164,
85
+ "Other - final_conv": 87
86
+ },
87
+ "latent_dimensions": {
88
+ "Level_0": {
89
+ "channels": 28,
90
+ "height": 96,
91
+ "width": 96,
92
+ "total_elements": 258048
93
+ },
94
+ "Level_1": {
95
+ "channels": 56,
96
+ "height": 48,
97
+ "width": 48,
98
+ "total_elements": 129024
99
+ },
100
+ "Level_2": {
101
+ "channels": 56,
102
+ "height": 24,
103
+ "width": 24,
104
+ "total_elements": 32256
105
+ },
106
+ "Level_3": {
107
+ "channels": 84,
108
+ "height": 12,
109
+ "width": 12,
110
+ "total_elements": 12096
111
+ }
112
+ },
113
+ "skip_dimensions": {
114
+ "Skip_0": {
115
+ "channels": 28,
116
+ "height": 96,
117
+ "width": 96,
118
+ "total_elements": 258048
119
+ },
120
+ "Skip_1": {
121
+ "channels": 56,
122
+ "height": 48,
123
+ "width": 48,
124
+ "total_elements": 129024
125
+ },
126
+ "Skip_2": {
127
+ "channels": 56,
128
+ "height": 24,
129
+ "width": 24,
130
+ "total_elements": 32256
131
+ }
132
+ },
133
+ "memory_analysis": {
134
+ "peak_memory_inference_mb": 1.927001953125,
135
+ "peak_memory_training_mb": 3.85400390625,
136
+ "peak_elements": 505152
137
+ }
138
+ },
139
+ "timestamp": "2025-11-06T18:10:09.761818"
140
+ }
exp_1762449009_9308_s96_f28_d3_m1_2_2_3/pipeline_results.json ADDED
@@ -0,0 +1,140 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "experiment_id": "exp_1762449009_8812_s96_f28_d3_m1_2_2_3",
3
+ "config": {
4
+ "input_size": 96,
5
+ "base_filters": 28,
6
+ "depth": 3,
7
+ "channel_multipliers": [
8
+ 1,
9
+ 2,
10
+ 2,
11
+ 3
12
+ ],
13
+ "n_channels": 8,
14
+ "n_classes": 3
15
+ },
16
+ "start_time": "2025-11-06T18:10:09.754685",
17
+ "success": true,
18
+ "timings": {
19
+ "onnx_conversion": 0.2888925038278103,
20
+ "openvino_conversion": 4.100632845889777,
21
+ "inference_total": 2.8084738550242037,
22
+ "inference_actual": 0.203952,
23
+ "total": 7.21151207992807
24
+ },
25
+ "errors": [],
26
+ "architecture_stats": {
27
+ "input_dimension": 73728,
28
+ "output_dimension": 27648,
29
+ "parameter_distribution": {
30
+ "encoder_params": 66304,
31
+ "decoder_params": 134540,
32
+ "bottleneck_params": 66024,
33
+ "other_params": 87,
34
+ "total_params": 266955,
35
+ "encoder_percentage": 24.837144837144837,
36
+ "decoder_percentage": 50.39800715476391,
37
+ "bottleneck_percentage": 24.73225824577176,
38
+ "other_percentage": 0.03258976231949205
39
+ },
40
+ "component_breakdown": {
41
+ "Encoder - encoders.0.channel_proj": 252,
42
+ "Encoder - encoders.0.convnext_block.dwconv": 1400,
43
+ "Encoder - encoders.0.convnext_block.norm": 56,
44
+ "Encoder - encoders.0.convnext_block.pwconv1": 3248,
45
+ "Encoder - encoders.0.convnext_block.pwconv2": 3164,
46
+ "Encoder - encoders.1.channel_proj": 1624,
47
+ "Encoder - encoders.1.convnext_block.dwconv": 2800,
48
+ "Encoder - encoders.1.convnext_block.norm": 112,
49
+ "Encoder - encoders.1.convnext_block.pwconv1": 12768,
50
+ "Encoder - encoders.1.convnext_block.pwconv2": 12600,
51
+ "Encoder - encoders.2.convnext_block.dwconv": 2800,
52
+ "Encoder - encoders.2.convnext_block.norm": 112,
53
+ "Encoder - encoders.2.convnext_block.pwconv1": 12768,
54
+ "Encoder - encoders.2.convnext_block.pwconv2": 12600,
55
+ "Bottleneck - bottleneck.channel_proj": 4788,
56
+ "Bottleneck - bottleneck.convnext_block.dwconv": 4200,
57
+ "Bottleneck - bottleneck.convnext_block.norm": 168,
58
+ "Bottleneck - bottleneck.convnext_block.pwconv1": 28560,
59
+ "Bottleneck - bottleneck.convnext_block.pwconv2": 28308,
60
+ "Decoder - upsamplers.0": 28308,
61
+ "Decoder - upsamplers.1": 12600,
62
+ "Decoder - upsamplers.2": 12600,
63
+ "Decoder - decoders.0.channel_proj": 7896,
64
+ "Decoder - decoders.0.convnext_block.dwconv": 2800,
65
+ "Decoder - decoders.0.convnext_block.norm": 112,
66
+ "Decoder - decoders.0.convnext_block.pwconv1": 12768,
67
+ "Decoder - decoders.0.convnext_block.pwconv2": 12600,
68
+ "Decoder - decoders.1.channel_proj": 6328,
69
+ "Decoder - decoders.1.convnext_block.dwconv": 2800,
70
+ "Decoder - decoders.1.convnext_block.norm": 112,
71
+ "Decoder - decoders.1.convnext_block.pwconv1": 12768,
72
+ "Decoder - decoders.1.convnext_block.pwconv2": 12600,
73
+ "Decoder - decoders.2.channel_proj": 2380,
74
+ "Decoder - decoders.2.convnext_block.dwconv": 1400,
75
+ "Decoder - decoders.2.convnext_block.norm": 56,
76
+ "Decoder - decoders.2.convnext_block.pwconv1": 3248,
77
+ "Decoder - decoders.2.convnext_block.pwconv2": 3164,
78
+ "Other - final_conv": 87
79
+ },
80
+ "latent_dimensions": {
81
+ "Level_0": {
82
+ "channels": 28,
83
+ "height": 96,
84
+ "width": 96,
85
+ "total_elements": 258048
86
+ },
87
+ "Level_1": {
88
+ "channels": 56,
89
+ "height": 48,
90
+ "width": 48,
91
+ "total_elements": 129024
92
+ },
93
+ "Level_2": {
94
+ "channels": 56,
95
+ "height": 24,
96
+ "width": 24,
97
+ "total_elements": 32256
98
+ },
99
+ "Level_3": {
100
+ "channels": 84,
101
+ "height": 12,
102
+ "width": 12,
103
+ "total_elements": 12096
104
+ }
105
+ },
106
+ "skip_dimensions": {
107
+ "Skip_0": {
108
+ "channels": 28,
109
+ "height": 96,
110
+ "width": 96,
111
+ "total_elements": 258048
112
+ },
113
+ "Skip_1": {
114
+ "channels": 56,
115
+ "height": 48,
116
+ "width": 48,
117
+ "total_elements": 129024
118
+ },
119
+ "Skip_2": {
120
+ "channels": 56,
121
+ "height": 24,
122
+ "width": 24,
123
+ "total_elements": 32256
124
+ }
125
+ },
126
+ "memory_analysis": {
127
+ "peak_memory_inference_mb": 1.927001953125,
128
+ "peak_memory_training_mb": 3.85400390625,
129
+ "peak_elements": 505152
130
+ }
131
+ },
132
+ "inference_results": {
133
+ "success": true,
134
+ "total_time": 2.8084738550242037,
135
+ "inference_time": 0.203952,
136
+ "stdout": "[setupvars.sh] OpenVINO environment initialized\nStarting Myriad inference...\nInput: /home/mount/experiments/exp_1762449009_9308_s96_f28_d3_m1_2_2_3/onnx/sample_input.npy\nModel XML: /home/mount/experiments/exp_1762449009_9308_s96_f28_d3_m1_2_2_3/openvino/model.xml\nModel BIN: /home/mount/experiments/exp_1762449009_9308_s96_f28_d3_m1_2_2_3/openvino/model.bin\n[OK] OpenVINO inference engine imported successfully\nDevice: MYRIAD\nModel XML: /home/mount/experiments/exp_1762449009_9308_s96_f28_d3_m1_2_2_3/openvino/model.xml\nModel BIN: /home/mount/experiments/exp_1762449009_9308_s96_f28_d3_m1_2_2_3/openvino/model.bin\nInput file: /home/mount/experiments/exp_1762449009_9308_s96_f28_d3_m1_2_2_3/onnx/sample_input.npy\nInitializing OpenVINO Runtime Core...\nAvailable devices:\n[E:] [BSL] found 0 ioexpander device\n ['CPU', 'GNA', 'MYRIAD']\nLoading network...\nInput blob: input\nInput shape: [1, 8, 96, 96]\nOutput blob: Conv_106\nOutput shape: [1, 3, 96, 96]\nLoading network to MYRIAD...\nLoading input data...\nInput data shape: (1, 8, 96, 96)\nInput data type: float32\nRunning inference...\n[OK] Inference completed!\nInference time: 0.203952 seconds\nOutput shape: (1, 3, 96, 96)\nOutput dtype: float32\nOutput range: [-0.605957, 0.702148]\nMyriad inference completed!\n",
137
+ "stderr": ""
138
+ },
139
+ "end_time": "2025-11-06T18:10:16.966186"
140
+ }
exp_1762449009_9308_s96_f28_d3_m1_2_2_3/run_inference.sh ADDED
@@ -0,0 +1,19 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #!/bin/bash
2
+ # Source OpenVINO environment
3
+ source /opt/intel/openvino/bin/setupvars.sh
4
+
5
+ PYPATH=/home/mount/scripts/simple_inference.py
6
+
7
+ echo "Starting Myriad inference..."
8
+ echo "Input: /home/mount/experiments/exp_1762449009_9308_s96_f28_d3_m1_2_2_3/onnx/sample_input.npy"
9
+ echo "Model XML: /home/mount/experiments/exp_1762449009_9308_s96_f28_d3_m1_2_2_3/openvino/model.xml"
10
+ echo "Model BIN: /home/mount/experiments/exp_1762449009_9308_s96_f28_d3_m1_2_2_3/openvino/model.bin"
11
+
12
+ python3 $PYPATH \
13
+ --input_filepath /home/mount/experiments/exp_1762449009_9308_s96_f28_d3_m1_2_2_3/onnx/sample_input.npy \
14
+ --device_name MYRIAD \
15
+ --model_bin /home/mount/experiments/exp_1762449009_9308_s96_f28_d3_m1_2_2_3/openvino/model.bin \
16
+ --model_xml /home/mount/experiments/exp_1762449009_9308_s96_f28_d3_m1_2_2_3/openvino/model.xml \
17
+ --output /home/mount/experiments/exp_1762449009_9308_s96_f28_d3_m1_2_2_3/inference/
18
+
19
+ echo "Myriad inference completed!"
exp_1762449060_9675_s96_f30_d3_m1_2_2_3/convert_openvino.sh ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #!/bin/bash
2
+ # Source OpenVINO environment
3
+ source /opt/intel/openvino/bin/setupvars.sh
4
+
5
+ # OpenVINO Model Optimizer script for ONNX model
6
+ PYPATH=/opt/intel/openvino_2020.3.194/deployment_tools/model_optimizer/mo.py
7
+ ONNX_MODEL=/home/mount/experiments/exp_1762449060_9675_s96_f30_d3_m1_2_2_3/onnx/model.onnx
8
+ OUTPUT_DIR=/home/mount/experiments/exp_1762449060_9675_s96_f30_d3_m1_2_2_3/openvino
9
+
10
+ echo "Converting ONNX model to OpenVINO IR format..."
11
+ echo "Input model: $ONNX_MODEL"
12
+ echo "Output directory: $OUTPUT_DIR"
13
+
14
+ # Create output directory if it doesn't exist
15
+ mkdir -p "$OUTPUT_DIR"
16
+
17
+ python3 $PYPATH \
18
+ --input_model "$ONNX_MODEL" \
19
+ --data_type FP16 \
20
+ --input_shape "[1,8,96,96]" \
21
+ --mean_values "[0,0,0,0,0,0,0,0]" \
22
+ --scale_values "[1,1,1,1,1,1,1,1]" \
23
+ --progress \
24
+ --stream_output \
25
+ --output_dir "$OUTPUT_DIR" \
26
+ --model_name model
27
+
28
+ echo "OpenVINO conversion completed!"
29
+ echo "Generated files in $OUTPUT_DIR:"
30
+ ls -la "$OUTPUT_DIR/"
exp_1762449060_9675_s96_f30_d3_m1_2_2_3/experiment.log ADDED
@@ -0,0 +1,100 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2025-11-06 18:11:00 - INFO - === Model Creation Phase ===
2
+ 2025-11-06 18:11:00 - INFO - Creating UNet model with config: {'input_size': 96, 'base_filters': 30, 'depth': 3, 'channel_multipliers': [1, 2, 2, 3], 'n_channels': 8, 'n_classes': 3}
3
+ 2025-11-06 18:11:00 - INFO - Model created successfully:
4
+ 2025-11-06 18:11:00 - INFO - Depth: 3
5
+ 2025-11-06 18:11:00 - INFO - Channels per level: [30, 60, 60, 90]
6
+ 2025-11-06 18:11:00 - INFO - Total parameters: 305,193
7
+ 2025-11-06 18:11:00 - INFO - Parameter memory: 1.16 MB
8
+ 2025-11-06 18:11:00 - INFO - === Detailed Architecture Analysis ===
9
+ 2025-11-06 18:11:00 - INFO - Input vector dimension: 73,728 (8 × 96 × 96)
10
+ 2025-11-06 18:11:00 - INFO - Component-wise parameter breakdown:
11
+ 2025-11-06 18:11:00 - INFO - Encoder - encoders.0.channel_proj: 270 parameters
12
+ 2025-11-06 18:11:00 - INFO - Encoder - encoders.0.convnext_block.dwconv: 1,500 parameters
13
+ 2025-11-06 18:11:00 - INFO - Encoder - encoders.0.convnext_block.norm: 60 parameters
14
+ 2025-11-06 18:11:00 - INFO - Encoder - encoders.0.convnext_block.pwconv1: 3,720 parameters
15
+ 2025-11-06 18:11:00 - INFO - Encoder - encoders.0.convnext_block.pwconv2: 3,630 parameters
16
+ 2025-11-06 18:11:00 - INFO - Encoder - encoders.1.channel_proj: 1,860 parameters
17
+ 2025-11-06 18:11:00 - INFO - Encoder - encoders.1.convnext_block.dwconv: 3,000 parameters
18
+ 2025-11-06 18:11:00 - INFO - Encoder - encoders.1.convnext_block.norm: 120 parameters
19
+ 2025-11-06 18:11:00 - INFO - Encoder - encoders.1.convnext_block.pwconv1: 14,640 parameters
20
+ 2025-11-06 18:11:00 - INFO - Encoder - encoders.1.convnext_block.pwconv2: 14,460 parameters
21
+ 2025-11-06 18:11:00 - INFO - Encoder - encoders.2.convnext_block.dwconv: 3,000 parameters
22
+ 2025-11-06 18:11:00 - INFO - Encoder - encoders.2.convnext_block.norm: 120 parameters
23
+ 2025-11-06 18:11:00 - INFO - Encoder - encoders.2.convnext_block.pwconv1: 14,640 parameters
24
+ 2025-11-06 18:11:00 - INFO - Encoder - encoders.2.convnext_block.pwconv2: 14,460 parameters
25
+ 2025-11-06 18:11:00 - INFO - Bottleneck - bottleneck.channel_proj: 5,490 parameters
26
+ 2025-11-06 18:11:00 - INFO - Bottleneck - bottleneck.convnext_block.dwconv: 4,500 parameters
27
+ 2025-11-06 18:11:00 - INFO - Bottleneck - bottleneck.convnext_block.norm: 180 parameters
28
+ 2025-11-06 18:11:00 - INFO - Bottleneck - bottleneck.convnext_block.pwconv1: 32,760 parameters
29
+ 2025-11-06 18:11:00 - INFO - Bottleneck - bottleneck.convnext_block.pwconv2: 32,490 parameters
30
+ 2025-11-06 18:11:00 - INFO - Decoder - upsamplers.0: 32,490 parameters
31
+ 2025-11-06 18:11:00 - INFO - Decoder - upsamplers.1: 14,460 parameters
32
+ 2025-11-06 18:11:00 - INFO - Decoder - upsamplers.2: 14,460 parameters
33
+ 2025-11-06 18:11:00 - INFO - Decoder - decoders.0.channel_proj: 9,060 parameters
34
+ 2025-11-06 18:11:00 - INFO - Decoder - decoders.0.convnext_block.dwconv: 3,000 parameters
35
+ 2025-11-06 18:11:00 - INFO - Decoder - decoders.0.convnext_block.norm: 120 parameters
36
+ 2025-11-06 18:11:00 - INFO - Decoder - decoders.0.convnext_block.pwconv1: 14,640 parameters
37
+ 2025-11-06 18:11:00 - INFO - Decoder - decoders.0.convnext_block.pwconv2: 14,460 parameters
38
+ 2025-11-06 18:11:00 - INFO - Decoder - decoders.1.channel_proj: 7,260 parameters
39
+ 2025-11-06 18:11:00 - INFO - Decoder - decoders.1.convnext_block.dwconv: 3,000 parameters
40
+ 2025-11-06 18:11:00 - INFO - Decoder - decoders.1.convnext_block.norm: 120 parameters
41
+ 2025-11-06 18:11:00 - INFO - Decoder - decoders.1.convnext_block.pwconv1: 14,640 parameters
42
+ 2025-11-06 18:11:00 - INFO - Decoder - decoders.1.convnext_block.pwconv2: 14,460 parameters
43
+ 2025-11-06 18:11:00 - INFO - Decoder - decoders.2.channel_proj: 2,730 parameters
44
+ 2025-11-06 18:11:00 - INFO - Decoder - decoders.2.convnext_block.dwconv: 1,500 parameters
45
+ 2025-11-06 18:11:00 - INFO - Decoder - decoders.2.convnext_block.norm: 60 parameters
46
+ 2025-11-06 18:11:00 - INFO - Decoder - decoders.2.convnext_block.pwconv1: 3,720 parameters
47
+ 2025-11-06 18:11:00 - INFO - Decoder - decoders.2.convnext_block.pwconv2: 3,630 parameters
48
+ 2025-11-06 18:11:00 - INFO - Other - final_conv: 93 parameters
49
+ 2025-11-06 18:11:00 - INFO - Parameter distribution summary:
50
+ 2025-11-06 18:11:00 - INFO - Encoder parameters: 75,480 (24.8%)
51
+ 2025-11-06 18:11:00 - INFO - Decoder parameters: 153,810 (50.5%)
52
+ 2025-11-06 18:11:00 - INFO - Bottleneck parameters: 75,420 (24.7%)
53
+ 2025-11-06 18:11:00 - INFO - Other parameters: 93 (0.0%)
54
+ 2025-11-06 18:11:00 - INFO - Latent space dimensions (feature maps at each level):
55
+ 2025-11-06 18:11:00 - INFO - Level 0: 30 × 96 × 96 = 276,480 elements
56
+ 2025-11-06 18:11:00 - INFO - Level 1: 60 × 48 × 48 = 138,240 elements
57
+ 2025-11-06 18:11:00 - INFO - Level 2: 60 × 24 × 24 = 34,560 elements
58
+ 2025-11-06 18:11:00 - INFO - Level 3: 90 × 12 × 12 = 12,960 elements
59
+ 2025-11-06 18:11:00 - INFO - Skip connection dimensions:
60
+ 2025-11-06 18:11:00 - INFO - Skip 0: 30 × 96 × 96 = 276,480 elements
61
+ 2025-11-06 18:11:00 - INFO - Skip 1: 60 × 48 × 48 = 138,240 elements
62
+ 2025-11-06 18:11:00 - INFO - Skip 2: 60 × 24 × 24 = 34,560 elements
63
+ 2025-11-06 18:11:00 - INFO - Memory analysis:
64
+ 2025-11-06 18:11:00 - INFO - Peak feature map memory (inference): 2.04 MB
65
+ 2025-11-06 18:11:00 - INFO - Peak feature map memory (training): 4.09 MB (with gradients)
66
+ 2025-11-06 18:11:00 - INFO - Output vector dimension: 27,648 (3 × 96 × 96)
67
+ 2025-11-06 18:11:00 - INFO - PyTorch model saved to: /home/philab/Desktop/hydranet/experiments/exp_1762449060_9675_s96_f30_d3_m1_2_2_3/pytorch/model.pt
68
+ 2025-11-06 18:11:00 - INFO - === ONNX Conversion Phase ===
69
+ 2025-11-06 18:11:00 - INFO - === Model Export Diagnostics ===
70
+ 2025-11-06 18:11:00 - INFO - PyTorch version: 1.9.0+cu102
71
+ 2025-11-06 18:11:00 - INFO - Model parameters: 305,193
72
+ 2025-11-06 18:11:00 - INFO - Model memory: 1.16 MB
73
+ 2025-11-06 18:11:00 - INFO - Starting ONNX export with opset version 11
74
+ 2025-11-06 18:11:00 - INFO - Model input shape: torch.Size([1, 8, 96, 96])
75
+ 2025-11-06 18:11:00 - INFO - Model input dtype: torch.float32
76
+ 2025-11-06 18:11:00 - INFO - Forward pass successful. Output shape: torch.Size([1, 3, 96, 96])
77
+ 2025-11-06 18:11:00 - INFO - Output dtype: torch.float32
78
+ 2025-11-06 18:11:00 - INFO - Output value range: [-0.6438, 0.5599]
79
+ 2025-11-06 18:11:00 - INFO - Model successfully exported to /home/philab/Desktop/hydranet/experiments/exp_1762449060_9675_s96_f30_d3_m1_2_2_3/onnx/model.onnx
80
+ 2025-11-06 18:11:00 - INFO - ONNX model size: 1.17 MB
81
+ 2025-11-06 18:11:00 - INFO - Saved dummy input with shape (1, 8, 96, 96) to /home/philab/Desktop/hydranet/experiments/exp_1762449060_9675_s96_f30_d3_m1_2_2_3/onnx/sample_input.npy
82
+ 2025-11-06 18:11:00 - INFO - Input data type: float32
83
+ 2025-11-06 18:11:00 - INFO - Input value range: [-4.3257, 4.2415]
84
+ 2025-11-06 18:11:00 - INFO - === OpenVINO Conversion Phase ===
85
+ 2025-11-06 18:11:00 - INFO - Starting OpenVINO conversion in Docker container...
86
+ 2025-11-06 18:11:04 - INFO - OpenVINO conversion completed in 4.08 seconds
87
+ 2025-11-06 18:11:04 - INFO - OpenVINO model files created:
88
+ 2025-11-06 18:11:04 - INFO - XML file: /home/philab/Desktop/hydranet/experiments/exp_1762449060_9675_s96_f30_d3_m1_2_2_3/openvino/model.xml (0.09 MB)
89
+ 2025-11-06 18:11:04 - INFO - BIN file: /home/philab/Desktop/hydranet/experiments/exp_1762449060_9675_s96_f30_d3_m1_2_2_3/openvino/model.bin (0.58 MB)
90
+ 2025-11-06 18:11:04 - INFO - === Myriad Inference Phase ===
91
+ 2025-11-06 18:11:04 - INFO - Starting Myriad inference in Docker container...
92
+ 2025-11-06 18:11:07 - INFO - Myriad inference completed in 2.90 seconds
93
+ 2025-11-06 18:11:07 - INFO - Actual inference time: 0.298507 seconds
94
+ 2025-11-06 18:11:07 - INFO - ✅ Complete pipeline executed successfully!
95
+ 2025-11-06 18:11:07 - INFO - ✅ Experiment 81 completed successfully
96
+ 2025-11-06 18:11:07 - INFO - Inference time: 0.298507s
97
+ 2025-11-06 18:11:07 - INFO -
98
+ === Experiment 82/2475 ===
99
+ 2025-11-06 18:11:07 - INFO - Experiment ID: exp_1762449067_7726_s96_f30_d3_m2_2_2_3
100
+ 2025-11-06 18:11:07 - INFO - Experiment directory: /home/philab/Desktop/hydranet/experiments/exp_1762449067_7726_s96_f30_d3_m2_2_2_3
exp_1762449060_9675_s96_f30_d3_m1_2_2_3/model_info.json ADDED
@@ -0,0 +1,140 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "config": {
3
+ "input_size": 96,
4
+ "base_filters": 30,
5
+ "depth": 3,
6
+ "channel_multipliers": [
7
+ 1,
8
+ 2,
9
+ 2,
10
+ 3
11
+ ],
12
+ "n_channels": 8,
13
+ "n_classes": 3
14
+ },
15
+ "model_info": {
16
+ "depth": 3,
17
+ "channels_per_level": [
18
+ 30,
19
+ 60,
20
+ 60,
21
+ 90
22
+ ],
23
+ "channel_multipliers": [
24
+ 1,
25
+ 2,
26
+ 2,
27
+ 3
28
+ ],
29
+ "total_parameters": 305193,
30
+ "trainable_parameters": 305193,
31
+ "model_size_mb": 1.1642189025878906
32
+ },
33
+ "architecture_stats": {
34
+ "input_dimension": 73728,
35
+ "output_dimension": 27648,
36
+ "parameter_distribution": {
37
+ "encoder_params": 75480,
38
+ "decoder_params": 153810,
39
+ "bottleneck_params": 75420,
40
+ "other_params": 93,
41
+ "total_params": 304803,
42
+ "encoder_percentage": 24.76353579197055,
43
+ "decoder_percentage": 50.462101750967015,
44
+ "bottleneck_percentage": 24.743850946348953,
45
+ "other_percentage": 0.03051151071347723
46
+ },
47
+ "component_breakdown": {
48
+ "Encoder - encoders.0.channel_proj": 270,
49
+ "Encoder - encoders.0.convnext_block.dwconv": 1500,
50
+ "Encoder - encoders.0.convnext_block.norm": 60,
51
+ "Encoder - encoders.0.convnext_block.pwconv1": 3720,
52
+ "Encoder - encoders.0.convnext_block.pwconv2": 3630,
53
+ "Encoder - encoders.1.channel_proj": 1860,
54
+ "Encoder - encoders.1.convnext_block.dwconv": 3000,
55
+ "Encoder - encoders.1.convnext_block.norm": 120,
56
+ "Encoder - encoders.1.convnext_block.pwconv1": 14640,
57
+ "Encoder - encoders.1.convnext_block.pwconv2": 14460,
58
+ "Encoder - encoders.2.convnext_block.dwconv": 3000,
59
+ "Encoder - encoders.2.convnext_block.norm": 120,
60
+ "Encoder - encoders.2.convnext_block.pwconv1": 14640,
61
+ "Encoder - encoders.2.convnext_block.pwconv2": 14460,
62
+ "Bottleneck - bottleneck.channel_proj": 5490,
63
+ "Bottleneck - bottleneck.convnext_block.dwconv": 4500,
64
+ "Bottleneck - bottleneck.convnext_block.norm": 180,
65
+ "Bottleneck - bottleneck.convnext_block.pwconv1": 32760,
66
+ "Bottleneck - bottleneck.convnext_block.pwconv2": 32490,
67
+ "Decoder - upsamplers.0": 32490,
68
+ "Decoder - upsamplers.1": 14460,
69
+ "Decoder - upsamplers.2": 14460,
70
+ "Decoder - decoders.0.channel_proj": 9060,
71
+ "Decoder - decoders.0.convnext_block.dwconv": 3000,
72
+ "Decoder - decoders.0.convnext_block.norm": 120,
73
+ "Decoder - decoders.0.convnext_block.pwconv1": 14640,
74
+ "Decoder - decoders.0.convnext_block.pwconv2": 14460,
75
+ "Decoder - decoders.1.channel_proj": 7260,
76
+ "Decoder - decoders.1.convnext_block.dwconv": 3000,
77
+ "Decoder - decoders.1.convnext_block.norm": 120,
78
+ "Decoder - decoders.1.convnext_block.pwconv1": 14640,
79
+ "Decoder - decoders.1.convnext_block.pwconv2": 14460,
80
+ "Decoder - decoders.2.channel_proj": 2730,
81
+ "Decoder - decoders.2.convnext_block.dwconv": 1500,
82
+ "Decoder - decoders.2.convnext_block.norm": 60,
83
+ "Decoder - decoders.2.convnext_block.pwconv1": 3720,
84
+ "Decoder - decoders.2.convnext_block.pwconv2": 3630,
85
+ "Other - final_conv": 93
86
+ },
87
+ "latent_dimensions": {
88
+ "Level_0": {
89
+ "channels": 30,
90
+ "height": 96,
91
+ "width": 96,
92
+ "total_elements": 276480
93
+ },
94
+ "Level_1": {
95
+ "channels": 60,
96
+ "height": 48,
97
+ "width": 48,
98
+ "total_elements": 138240
99
+ },
100
+ "Level_2": {
101
+ "channels": 60,
102
+ "height": 24,
103
+ "width": 24,
104
+ "total_elements": 34560
105
+ },
106
+ "Level_3": {
107
+ "channels": 90,
108
+ "height": 12,
109
+ "width": 12,
110
+ "total_elements": 12960
111
+ }
112
+ },
113
+ "skip_dimensions": {
114
+ "Skip_0": {
115
+ "channels": 30,
116
+ "height": 96,
117
+ "width": 96,
118
+ "total_elements": 276480
119
+ },
120
+ "Skip_1": {
121
+ "channels": 60,
122
+ "height": 48,
123
+ "width": 48,
124
+ "total_elements": 138240
125
+ },
126
+ "Skip_2": {
127
+ "channels": 60,
128
+ "height": 24,
129
+ "width": 24,
130
+ "total_elements": 34560
131
+ }
132
+ },
133
+ "memory_analysis": {
134
+ "peak_memory_inference_mb": 2.0445556640625,
135
+ "peak_memory_training_mb": 4.089111328125,
136
+ "peak_elements": 535968
137
+ }
138
+ },
139
+ "timestamp": "2025-11-06T18:11:00.552144"
140
+ }
exp_1762449060_9675_s96_f30_d3_m1_2_2_3/pipeline_results.json ADDED
@@ -0,0 +1,140 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "experiment_id": "exp_1762449060_7398_s96_f30_d3_m1_2_2_3",
3
+ "config": {
4
+ "input_size": 96,
5
+ "base_filters": 30,
6
+ "depth": 3,
7
+ "channel_multipliers": [
8
+ 1,
9
+ 2,
10
+ 2,
11
+ 3
12
+ ],
13
+ "n_channels": 8,
14
+ "n_classes": 3
15
+ },
16
+ "start_time": "2025-11-06T18:11:00.543091",
17
+ "success": true,
18
+ "timings": {
19
+ "onnx_conversion": 0.295903658028692,
20
+ "openvino_conversion": 4.082625380018726,
21
+ "inference_total": 2.8962551350705326,
22
+ "inference_actual": 0.298507,
23
+ "total": 7.290062549058348
24
+ },
25
+ "errors": [],
26
+ "architecture_stats": {
27
+ "input_dimension": 73728,
28
+ "output_dimension": 27648,
29
+ "parameter_distribution": {
30
+ "encoder_params": 75480,
31
+ "decoder_params": 153810,
32
+ "bottleneck_params": 75420,
33
+ "other_params": 93,
34
+ "total_params": 304803,
35
+ "encoder_percentage": 24.76353579197055,
36
+ "decoder_percentage": 50.462101750967015,
37
+ "bottleneck_percentage": 24.743850946348953,
38
+ "other_percentage": 0.03051151071347723
39
+ },
40
+ "component_breakdown": {
41
+ "Encoder - encoders.0.channel_proj": 270,
42
+ "Encoder - encoders.0.convnext_block.dwconv": 1500,
43
+ "Encoder - encoders.0.convnext_block.norm": 60,
44
+ "Encoder - encoders.0.convnext_block.pwconv1": 3720,
45
+ "Encoder - encoders.0.convnext_block.pwconv2": 3630,
46
+ "Encoder - encoders.1.channel_proj": 1860,
47
+ "Encoder - encoders.1.convnext_block.dwconv": 3000,
48
+ "Encoder - encoders.1.convnext_block.norm": 120,
49
+ "Encoder - encoders.1.convnext_block.pwconv1": 14640,
50
+ "Encoder - encoders.1.convnext_block.pwconv2": 14460,
51
+ "Encoder - encoders.2.convnext_block.dwconv": 3000,
52
+ "Encoder - encoders.2.convnext_block.norm": 120,
53
+ "Encoder - encoders.2.convnext_block.pwconv1": 14640,
54
+ "Encoder - encoders.2.convnext_block.pwconv2": 14460,
55
+ "Bottleneck - bottleneck.channel_proj": 5490,
56
+ "Bottleneck - bottleneck.convnext_block.dwconv": 4500,
57
+ "Bottleneck - bottleneck.convnext_block.norm": 180,
58
+ "Bottleneck - bottleneck.convnext_block.pwconv1": 32760,
59
+ "Bottleneck - bottleneck.convnext_block.pwconv2": 32490,
60
+ "Decoder - upsamplers.0": 32490,
61
+ "Decoder - upsamplers.1": 14460,
62
+ "Decoder - upsamplers.2": 14460,
63
+ "Decoder - decoders.0.channel_proj": 9060,
64
+ "Decoder - decoders.0.convnext_block.dwconv": 3000,
65
+ "Decoder - decoders.0.convnext_block.norm": 120,
66
+ "Decoder - decoders.0.convnext_block.pwconv1": 14640,
67
+ "Decoder - decoders.0.convnext_block.pwconv2": 14460,
68
+ "Decoder - decoders.1.channel_proj": 7260,
69
+ "Decoder - decoders.1.convnext_block.dwconv": 3000,
70
+ "Decoder - decoders.1.convnext_block.norm": 120,
71
+ "Decoder - decoders.1.convnext_block.pwconv1": 14640,
72
+ "Decoder - decoders.1.convnext_block.pwconv2": 14460,
73
+ "Decoder - decoders.2.channel_proj": 2730,
74
+ "Decoder - decoders.2.convnext_block.dwconv": 1500,
75
+ "Decoder - decoders.2.convnext_block.norm": 60,
76
+ "Decoder - decoders.2.convnext_block.pwconv1": 3720,
77
+ "Decoder - decoders.2.convnext_block.pwconv2": 3630,
78
+ "Other - final_conv": 93
79
+ },
80
+ "latent_dimensions": {
81
+ "Level_0": {
82
+ "channels": 30,
83
+ "height": 96,
84
+ "width": 96,
85
+ "total_elements": 276480
86
+ },
87
+ "Level_1": {
88
+ "channels": 60,
89
+ "height": 48,
90
+ "width": 48,
91
+ "total_elements": 138240
92
+ },
93
+ "Level_2": {
94
+ "channels": 60,
95
+ "height": 24,
96
+ "width": 24,
97
+ "total_elements": 34560
98
+ },
99
+ "Level_3": {
100
+ "channels": 90,
101
+ "height": 12,
102
+ "width": 12,
103
+ "total_elements": 12960
104
+ }
105
+ },
106
+ "skip_dimensions": {
107
+ "Skip_0": {
108
+ "channels": 30,
109
+ "height": 96,
110
+ "width": 96,
111
+ "total_elements": 276480
112
+ },
113
+ "Skip_1": {
114
+ "channels": 60,
115
+ "height": 48,
116
+ "width": 48,
117
+ "total_elements": 138240
118
+ },
119
+ "Skip_2": {
120
+ "channels": 60,
121
+ "height": 24,
122
+ "width": 24,
123
+ "total_elements": 34560
124
+ }
125
+ },
126
+ "memory_analysis": {
127
+ "peak_memory_inference_mb": 2.0445556640625,
128
+ "peak_memory_training_mb": 4.089111328125,
129
+ "peak_elements": 535968
130
+ }
131
+ },
132
+ "inference_results": {
133
+ "success": true,
134
+ "total_time": 2.8962551350705326,
135
+ "inference_time": 0.298507,
136
+ "stdout": "[setupvars.sh] OpenVINO environment initialized\nStarting Myriad inference...\nInput: /home/mount/experiments/exp_1762449060_9675_s96_f30_d3_m1_2_2_3/onnx/sample_input.npy\nModel XML: /home/mount/experiments/exp_1762449060_9675_s96_f30_d3_m1_2_2_3/openvino/model.xml\nModel BIN: /home/mount/experiments/exp_1762449060_9675_s96_f30_d3_m1_2_2_3/openvino/model.bin\n[OK] OpenVINO inference engine imported successfully\nDevice: MYRIAD\nModel XML: /home/mount/experiments/exp_1762449060_9675_s96_f30_d3_m1_2_2_3/openvino/model.xml\nModel BIN: /home/mount/experiments/exp_1762449060_9675_s96_f30_d3_m1_2_2_3/openvino/model.bin\nInput file: /home/mount/experiments/exp_1762449060_9675_s96_f30_d3_m1_2_2_3/onnx/sample_input.npy\nInitializing OpenVINO Runtime Core...\nAvailable devices:\n[E:] [BSL] found 0 ioexpander device\n ['CPU', 'GNA', 'MYRIAD']\nLoading network...\nInput blob: input\nInput shape: [1, 8, 96, 96]\nOutput blob: Conv_106\nOutput shape: [1, 3, 96, 96]\nLoading network to MYRIAD...\nLoading input data...\nInput data shape: (1, 8, 96, 96)\nInput data type: float32\nRunning inference...\n[OK] Inference completed!\nInference time: 0.298507 seconds\nOutput shape: (1, 3, 96, 96)\nOutput dtype: float32\nOutput range: [-0.643066, 0.559570]\nMyriad inference completed!\n",
137
+ "stderr": ""
138
+ },
139
+ "end_time": "2025-11-06T18:11:07.833143"
140
+ }
exp_1762449060_9675_s96_f30_d3_m1_2_2_3/run_inference.sh ADDED
@@ -0,0 +1,19 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #!/bin/bash
2
+ # Source OpenVINO environment
3
+ source /opt/intel/openvino/bin/setupvars.sh
4
+
5
+ PYPATH=/home/mount/scripts/simple_inference.py
6
+
7
+ echo "Starting Myriad inference..."
8
+ echo "Input: /home/mount/experiments/exp_1762449060_9675_s96_f30_d3_m1_2_2_3/onnx/sample_input.npy"
9
+ echo "Model XML: /home/mount/experiments/exp_1762449060_9675_s96_f30_d3_m1_2_2_3/openvino/model.xml"
10
+ echo "Model BIN: /home/mount/experiments/exp_1762449060_9675_s96_f30_d3_m1_2_2_3/openvino/model.bin"
11
+
12
+ python3 $PYPATH \
13
+ --input_filepath /home/mount/experiments/exp_1762449060_9675_s96_f30_d3_m1_2_2_3/onnx/sample_input.npy \
14
+ --device_name MYRIAD \
15
+ --model_bin /home/mount/experiments/exp_1762449060_9675_s96_f30_d3_m1_2_2_3/openvino/model.bin \
16
+ --model_xml /home/mount/experiments/exp_1762449060_9675_s96_f30_d3_m1_2_2_3/openvino/model.xml \
17
+ --output /home/mount/experiments/exp_1762449060_9675_s96_f30_d3_m1_2_2_3/inference/
18
+
19
+ echo "Myriad inference completed!"
exp_1762449133_5525_s96_f32_d3_m1_2_2_2/convert_openvino.sh ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #!/bin/bash
2
+ # Source OpenVINO environment
3
+ source /opt/intel/openvino/bin/setupvars.sh
4
+
5
+ # OpenVINO Model Optimizer script for ONNX model
6
+ PYPATH=/opt/intel/openvino_2020.3.194/deployment_tools/model_optimizer/mo.py
7
+ ONNX_MODEL=/home/mount/experiments/exp_1762449133_5525_s96_f32_d3_m1_2_2_2/onnx/model.onnx
8
+ OUTPUT_DIR=/home/mount/experiments/exp_1762449133_5525_s96_f32_d3_m1_2_2_2/openvino
9
+
10
+ echo "Converting ONNX model to OpenVINO IR format..."
11
+ echo "Input model: $ONNX_MODEL"
12
+ echo "Output directory: $OUTPUT_DIR"
13
+
14
+ # Create output directory if it doesn't exist
15
+ mkdir -p "$OUTPUT_DIR"
16
+
17
+ python3 $PYPATH \
18
+ --input_model "$ONNX_MODEL" \
19
+ --data_type FP16 \
20
+ --input_shape "[1,8,96,96]" \
21
+ --mean_values "[0,0,0,0,0,0,0,0]" \
22
+ --scale_values "[1,1,1,1,1,1,1,1]" \
23
+ --progress \
24
+ --stream_output \
25
+ --output_dir "$OUTPUT_DIR" \
26
+ --model_name model
27
+
28
+ echo "OpenVINO conversion completed!"
29
+ echo "Generated files in $OUTPUT_DIR:"
30
+ ls -la "$OUTPUT_DIR/"
exp_1762449133_5525_s96_f32_d3_m1_2_2_2/experiment.log ADDED
@@ -0,0 +1,99 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2025-11-06 18:12:13 - INFO - === Model Creation Phase ===
2
+ 2025-11-06 18:12:13 - INFO - Creating UNet model with config: {'input_size': 96, 'base_filters': 32, 'depth': 3, 'channel_multipliers': [1, 2, 2, 2], 'n_channels': 8, 'n_classes': 3}
3
+ 2025-11-06 18:12:13 - INFO - Model created successfully:
4
+ 2025-11-06 18:12:13 - INFO - Depth: 3
5
+ 2025-11-06 18:12:13 - INFO - Channels per level: [32, 64, 64, 64]
6
+ 2025-11-06 18:12:13 - INFO - Total parameters: 273,955
7
+ 2025-11-06 18:12:13 - INFO - Parameter memory: 1.05 MB
8
+ 2025-11-06 18:12:13 - INFO - === Detailed Architecture Analysis ===
9
+ 2025-11-06 18:12:13 - INFO - Input vector dimension: 73,728 (8 × 96 × 96)
10
+ 2025-11-06 18:12:13 - INFO - Component-wise parameter breakdown:
11
+ 2025-11-06 18:12:13 - INFO - Encoder - encoders.0.channel_proj: 288 parameters
12
+ 2025-11-06 18:12:13 - INFO - Encoder - encoders.0.convnext_block.dwconv: 1,600 parameters
13
+ 2025-11-06 18:12:13 - INFO - Encoder - encoders.0.convnext_block.norm: 64 parameters
14
+ 2025-11-06 18:12:13 - INFO - Encoder - encoders.0.convnext_block.pwconv1: 4,224 parameters
15
+ 2025-11-06 18:12:13 - INFO - Encoder - encoders.0.convnext_block.pwconv2: 4,128 parameters
16
+ 2025-11-06 18:12:13 - INFO - Encoder - encoders.1.channel_proj: 2,112 parameters
17
+ 2025-11-06 18:12:13 - INFO - Encoder - encoders.1.convnext_block.dwconv: 3,200 parameters
18
+ 2025-11-06 18:12:13 - INFO - Encoder - encoders.1.convnext_block.norm: 128 parameters
19
+ 2025-11-06 18:12:13 - INFO - Encoder - encoders.1.convnext_block.pwconv1: 16,640 parameters
20
+ 2025-11-06 18:12:13 - INFO - Encoder - encoders.1.convnext_block.pwconv2: 16,448 parameters
21
+ 2025-11-06 18:12:13 - INFO - Encoder - encoders.2.convnext_block.dwconv: 3,200 parameters
22
+ 2025-11-06 18:12:13 - INFO - Encoder - encoders.2.convnext_block.norm: 128 parameters
23
+ 2025-11-06 18:12:13 - INFO - Encoder - encoders.2.convnext_block.pwconv1: 16,640 parameters
24
+ 2025-11-06 18:12:13 - INFO - Encoder - encoders.2.convnext_block.pwconv2: 16,448 parameters
25
+ 2025-11-06 18:12:13 - INFO - Bottleneck - bottleneck.convnext_block.dwconv: 3,200 parameters
26
+ 2025-11-06 18:12:13 - INFO - Bottleneck - bottleneck.convnext_block.norm: 128 parameters
27
+ 2025-11-06 18:12:13 - INFO - Bottleneck - bottleneck.convnext_block.pwconv1: 16,640 parameters
28
+ 2025-11-06 18:12:13 - INFO - Bottleneck - bottleneck.convnext_block.pwconv2: 16,448 parameters
29
+ 2025-11-06 18:12:13 - INFO - Decoder - upsamplers.0: 16,448 parameters
30
+ 2025-11-06 18:12:13 - INFO - Decoder - upsamplers.1: 16,448 parameters
31
+ 2025-11-06 18:12:13 - INFO - Decoder - upsamplers.2: 16,448 parameters
32
+ 2025-11-06 18:12:13 - INFO - Decoder - decoders.0.channel_proj: 8,256 parameters
33
+ 2025-11-06 18:12:13 - INFO - Decoder - decoders.0.convnext_block.dwconv: 3,200 parameters
34
+ 2025-11-06 18:12:13 - INFO - Decoder - decoders.0.convnext_block.norm: 128 parameters
35
+ 2025-11-06 18:12:13 - INFO - Decoder - decoders.0.convnext_block.pwconv1: 16,640 parameters
36
+ 2025-11-06 18:12:13 - INFO - Decoder - decoders.0.convnext_block.pwconv2: 16,448 parameters
37
+ 2025-11-06 18:12:13 - INFO - Decoder - decoders.1.channel_proj: 8,256 parameters
38
+ 2025-11-06 18:12:13 - INFO - Decoder - decoders.1.convnext_block.dwconv: 3,200 parameters
39
+ 2025-11-06 18:12:13 - INFO - Decoder - decoders.1.convnext_block.norm: 128 parameters
40
+ 2025-11-06 18:12:13 - INFO - Decoder - decoders.1.convnext_block.pwconv1: 16,640 parameters
41
+ 2025-11-06 18:12:13 - INFO - Decoder - decoders.1.convnext_block.pwconv2: 16,448 parameters
42
+ 2025-11-06 18:12:13 - INFO - Decoder - decoders.2.channel_proj: 3,104 parameters
43
+ 2025-11-06 18:12:13 - INFO - Decoder - decoders.2.convnext_block.dwconv: 1,600 parameters
44
+ 2025-11-06 18:12:13 - INFO - Decoder - decoders.2.convnext_block.norm: 64 parameters
45
+ 2025-11-06 18:12:13 - INFO - Decoder - decoders.2.convnext_block.pwconv1: 4,224 parameters
46
+ 2025-11-06 18:12:13 - INFO - Decoder - decoders.2.convnext_block.pwconv2: 4,128 parameters
47
+ 2025-11-06 18:12:13 - INFO - Other - final_conv: 99 parameters
48
+ 2025-11-06 18:12:13 - INFO - Parameter distribution summary:
49
+ 2025-11-06 18:12:13 - INFO - Encoder parameters: 85,248 (31.2%)
50
+ 2025-11-06 18:12:13 - INFO - Decoder parameters: 151,808 (55.5%)
51
+ 2025-11-06 18:12:13 - INFO - Bottleneck parameters: 36,416 (13.3%)
52
+ 2025-11-06 18:12:13 - INFO - Other parameters: 99 (0.0%)
53
+ 2025-11-06 18:12:13 - INFO - Latent space dimensions (feature maps at each level):
54
+ 2025-11-06 18:12:13 - INFO - Level 0: 32 × 96 × 96 = 294,912 elements
55
+ 2025-11-06 18:12:13 - INFO - Level 1: 64 × 48 × 48 = 147,456 elements
56
+ 2025-11-06 18:12:13 - INFO - Level 2: 64 × 24 × 24 = 36,864 elements
57
+ 2025-11-06 18:12:13 - INFO - Level 3: 64 × 12 × 12 = 9,216 elements
58
+ 2025-11-06 18:12:13 - INFO - Skip connection dimensions:
59
+ 2025-11-06 18:12:13 - INFO - Skip 0: 32 × 96 × 96 = 294,912 elements
60
+ 2025-11-06 18:12:13 - INFO - Skip 1: 64 × 48 × 48 = 147,456 elements
61
+ 2025-11-06 18:12:13 - INFO - Skip 2: 64 × 24 × 24 = 36,864 elements
62
+ 2025-11-06 18:12:13 - INFO - Memory analysis:
63
+ 2025-11-06 18:12:13 - INFO - Peak feature map memory (inference): 2.14 MB
64
+ 2025-11-06 18:12:13 - INFO - Peak feature map memory (training): 4.29 MB (with gradients)
65
+ 2025-11-06 18:12:13 - INFO - Output vector dimension: 27,648 (3 × 96 × 96)
66
+ 2025-11-06 18:12:13 - INFO - PyTorch model saved to: /home/philab/Desktop/hydranet/experiments/exp_1762449133_5525_s96_f32_d3_m1_2_2_2/pytorch/model.pt
67
+ 2025-11-06 18:12:13 - INFO - === ONNX Conversion Phase ===
68
+ 2025-11-06 18:12:13 - INFO - === Model Export Diagnostics ===
69
+ 2025-11-06 18:12:13 - INFO - PyTorch version: 1.9.0+cu102
70
+ 2025-11-06 18:12:13 - INFO - Model parameters: 273,955
71
+ 2025-11-06 18:12:13 - INFO - Model memory: 1.05 MB
72
+ 2025-11-06 18:12:13 - INFO - Starting ONNX export with opset version 11
73
+ 2025-11-06 18:12:13 - INFO - Model input shape: torch.Size([1, 8, 96, 96])
74
+ 2025-11-06 18:12:13 - INFO - Model input dtype: torch.float32
75
+ 2025-11-06 18:12:13 - INFO - Forward pass successful. Output shape: torch.Size([1, 3, 96, 96])
76
+ 2025-11-06 18:12:13 - INFO - Output dtype: torch.float32
77
+ 2025-11-06 18:12:13 - INFO - Output value range: [-0.6360, 0.3016]
78
+ 2025-11-06 18:12:13 - INFO - Model successfully exported to /home/philab/Desktop/hydranet/experiments/exp_1762449133_5525_s96_f32_d3_m1_2_2_2/onnx/model.onnx
79
+ 2025-11-06 18:12:13 - INFO - ONNX model size: 1.05 MB
80
+ 2025-11-06 18:12:13 - INFO - Saved dummy input with shape (1, 8, 96, 96) to /home/philab/Desktop/hydranet/experiments/exp_1762449133_5525_s96_f32_d3_m1_2_2_2/onnx/sample_input.npy
81
+ 2025-11-06 18:12:13 - INFO - Input data type: float32
82
+ 2025-11-06 18:12:13 - INFO - Input value range: [-4.9373, 4.2185]
83
+ 2025-11-06 18:12:13 - INFO - === OpenVINO Conversion Phase ===
84
+ 2025-11-06 18:12:13 - INFO - Starting OpenVINO conversion in Docker container...
85
+ 2025-11-06 18:12:17 - INFO - OpenVINO conversion completed in 4.02 seconds
86
+ 2025-11-06 18:12:17 - INFO - OpenVINO model files created:
87
+ 2025-11-06 18:12:17 - INFO - XML file: /home/philab/Desktop/hydranet/experiments/exp_1762449133_5525_s96_f32_d3_m1_2_2_2/openvino/model.xml (0.08 MB)
88
+ 2025-11-06 18:12:17 - INFO - BIN file: /home/philab/Desktop/hydranet/experiments/exp_1762449133_5525_s96_f32_d3_m1_2_2_2/openvino/model.bin (0.52 MB)
89
+ 2025-11-06 18:12:17 - INFO - === Myriad Inference Phase ===
90
+ 2025-11-06 18:12:17 - INFO - Starting Myriad inference in Docker container...
91
+ 2025-11-06 18:12:20 - INFO - Myriad inference completed in 2.74 seconds
92
+ 2025-11-06 18:12:20 - INFO - Actual inference time: 0.145800 seconds
93
+ 2025-11-06 18:12:20 - INFO - ✅ Complete pipeline executed successfully!
94
+ 2025-11-06 18:12:20 - INFO - ✅ Experiment 91 completed successfully
95
+ 2025-11-06 18:12:20 - INFO - Inference time: 0.145800s
96
+ 2025-11-06 18:12:20 - INFO -
97
+ === Experiment 92/2475 ===
98
+ 2025-11-06 18:12:20 - INFO - Experiment ID: exp_1762449140_4143_s96_f32_d3_m1_2_2_3
99
+ 2025-11-06 18:12:20 - INFO - Experiment directory: /home/philab/Desktop/hydranet/experiments/exp_1762449140_4143_s96_f32_d3_m1_2_2_3
exp_1762449133_5525_s96_f32_d3_m1_2_2_2/model_info.json ADDED
@@ -0,0 +1,139 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "config": {
3
+ "input_size": 96,
4
+ "base_filters": 32,
5
+ "depth": 3,
6
+ "channel_multipliers": [
7
+ 1,
8
+ 2,
9
+ 2,
10
+ 2
11
+ ],
12
+ "n_channels": 8,
13
+ "n_classes": 3
14
+ },
15
+ "model_info": {
16
+ "depth": 3,
17
+ "channels_per_level": [
18
+ 32,
19
+ 64,
20
+ 64,
21
+ 64
22
+ ],
23
+ "channel_multipliers": [
24
+ 1,
25
+ 2,
26
+ 2,
27
+ 2
28
+ ],
29
+ "total_parameters": 273955,
30
+ "trainable_parameters": 273955,
31
+ "model_size_mb": 1.0450553894042969
32
+ },
33
+ "architecture_stats": {
34
+ "input_dimension": 73728,
35
+ "output_dimension": 27648,
36
+ "parameter_distribution": {
37
+ "encoder_params": 85248,
38
+ "decoder_params": 151808,
39
+ "bottleneck_params": 36416,
40
+ "other_params": 99,
41
+ "total_params": 273571,
42
+ "encoder_percentage": 31.161197641562882,
43
+ "decoder_percentage": 55.49126186620658,
44
+ "bottleneck_percentage": 13.311352445982944,
45
+ "other_percentage": 0.03618804624759203
46
+ },
47
+ "component_breakdown": {
48
+ "Encoder - encoders.0.channel_proj": 288,
49
+ "Encoder - encoders.0.convnext_block.dwconv": 1600,
50
+ "Encoder - encoders.0.convnext_block.norm": 64,
51
+ "Encoder - encoders.0.convnext_block.pwconv1": 4224,
52
+ "Encoder - encoders.0.convnext_block.pwconv2": 4128,
53
+ "Encoder - encoders.1.channel_proj": 2112,
54
+ "Encoder - encoders.1.convnext_block.dwconv": 3200,
55
+ "Encoder - encoders.1.convnext_block.norm": 128,
56
+ "Encoder - encoders.1.convnext_block.pwconv1": 16640,
57
+ "Encoder - encoders.1.convnext_block.pwconv2": 16448,
58
+ "Encoder - encoders.2.convnext_block.dwconv": 3200,
59
+ "Encoder - encoders.2.convnext_block.norm": 128,
60
+ "Encoder - encoders.2.convnext_block.pwconv1": 16640,
61
+ "Encoder - encoders.2.convnext_block.pwconv2": 16448,
62
+ "Bottleneck - bottleneck.convnext_block.dwconv": 3200,
63
+ "Bottleneck - bottleneck.convnext_block.norm": 128,
64
+ "Bottleneck - bottleneck.convnext_block.pwconv1": 16640,
65
+ "Bottleneck - bottleneck.convnext_block.pwconv2": 16448,
66
+ "Decoder - upsamplers.0": 16448,
67
+ "Decoder - upsamplers.1": 16448,
68
+ "Decoder - upsamplers.2": 16448,
69
+ "Decoder - decoders.0.channel_proj": 8256,
70
+ "Decoder - decoders.0.convnext_block.dwconv": 3200,
71
+ "Decoder - decoders.0.convnext_block.norm": 128,
72
+ "Decoder - decoders.0.convnext_block.pwconv1": 16640,
73
+ "Decoder - decoders.0.convnext_block.pwconv2": 16448,
74
+ "Decoder - decoders.1.channel_proj": 8256,
75
+ "Decoder - decoders.1.convnext_block.dwconv": 3200,
76
+ "Decoder - decoders.1.convnext_block.norm": 128,
77
+ "Decoder - decoders.1.convnext_block.pwconv1": 16640,
78
+ "Decoder - decoders.1.convnext_block.pwconv2": 16448,
79
+ "Decoder - decoders.2.channel_proj": 3104,
80
+ "Decoder - decoders.2.convnext_block.dwconv": 1600,
81
+ "Decoder - decoders.2.convnext_block.norm": 64,
82
+ "Decoder - decoders.2.convnext_block.pwconv1": 4224,
83
+ "Decoder - decoders.2.convnext_block.pwconv2": 4128,
84
+ "Other - final_conv": 99
85
+ },
86
+ "latent_dimensions": {
87
+ "Level_0": {
88
+ "channels": 32,
89
+ "height": 96,
90
+ "width": 96,
91
+ "total_elements": 294912
92
+ },
93
+ "Level_1": {
94
+ "channels": 64,
95
+ "height": 48,
96
+ "width": 48,
97
+ "total_elements": 147456
98
+ },
99
+ "Level_2": {
100
+ "channels": 64,
101
+ "height": 24,
102
+ "width": 24,
103
+ "total_elements": 36864
104
+ },
105
+ "Level_3": {
106
+ "channels": 64,
107
+ "height": 12,
108
+ "width": 12,
109
+ "total_elements": 9216
110
+ }
111
+ },
112
+ "skip_dimensions": {
113
+ "Skip_0": {
114
+ "channels": 32,
115
+ "height": 96,
116
+ "width": 96,
117
+ "total_elements": 294912
118
+ },
119
+ "Skip_1": {
120
+ "channels": 64,
121
+ "height": 48,
122
+ "width": 48,
123
+ "total_elements": 147456
124
+ },
125
+ "Skip_2": {
126
+ "channels": 64,
127
+ "height": 24,
128
+ "width": 24,
129
+ "total_elements": 36864
130
+ }
131
+ },
132
+ "memory_analysis": {
133
+ "peak_memory_inference_mb": 2.14453125,
134
+ "peak_memory_training_mb": 4.2890625,
135
+ "peak_elements": 562176
136
+ }
137
+ },
138
+ "timestamp": "2025-11-06T18:12:13.447999"
139
+ }
exp_1762449133_5525_s96_f32_d3_m1_2_2_2/pipeline_results.json ADDED
@@ -0,0 +1,139 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "experiment_id": "exp_1762449133_3385_s96_f32_d3_m1_2_2_2",
3
+ "config": {
4
+ "input_size": 96,
5
+ "base_filters": 32,
6
+ "depth": 3,
7
+ "channel_multipliers": [
8
+ 1,
9
+ 2,
10
+ 2,
11
+ 2
12
+ ],
13
+ "n_channels": 8,
14
+ "n_classes": 3
15
+ },
16
+ "start_time": "2025-11-06T18:12:13.440071",
17
+ "success": true,
18
+ "timings": {
19
+ "onnx_conversion": 0.2942611768376082,
20
+ "openvino_conversion": 4.017048191046342,
21
+ "inference_total": 2.742616229923442,
22
+ "inference_actual": 0.1458,
23
+ "total": 7.0675131909083575
24
+ },
25
+ "errors": [],
26
+ "architecture_stats": {
27
+ "input_dimension": 73728,
28
+ "output_dimension": 27648,
29
+ "parameter_distribution": {
30
+ "encoder_params": 85248,
31
+ "decoder_params": 151808,
32
+ "bottleneck_params": 36416,
33
+ "other_params": 99,
34
+ "total_params": 273571,
35
+ "encoder_percentage": 31.161197641562882,
36
+ "decoder_percentage": 55.49126186620658,
37
+ "bottleneck_percentage": 13.311352445982944,
38
+ "other_percentage": 0.03618804624759203
39
+ },
40
+ "component_breakdown": {
41
+ "Encoder - encoders.0.channel_proj": 288,
42
+ "Encoder - encoders.0.convnext_block.dwconv": 1600,
43
+ "Encoder - encoders.0.convnext_block.norm": 64,
44
+ "Encoder - encoders.0.convnext_block.pwconv1": 4224,
45
+ "Encoder - encoders.0.convnext_block.pwconv2": 4128,
46
+ "Encoder - encoders.1.channel_proj": 2112,
47
+ "Encoder - encoders.1.convnext_block.dwconv": 3200,
48
+ "Encoder - encoders.1.convnext_block.norm": 128,
49
+ "Encoder - encoders.1.convnext_block.pwconv1": 16640,
50
+ "Encoder - encoders.1.convnext_block.pwconv2": 16448,
51
+ "Encoder - encoders.2.convnext_block.dwconv": 3200,
52
+ "Encoder - encoders.2.convnext_block.norm": 128,
53
+ "Encoder - encoders.2.convnext_block.pwconv1": 16640,
54
+ "Encoder - encoders.2.convnext_block.pwconv2": 16448,
55
+ "Bottleneck - bottleneck.convnext_block.dwconv": 3200,
56
+ "Bottleneck - bottleneck.convnext_block.norm": 128,
57
+ "Bottleneck - bottleneck.convnext_block.pwconv1": 16640,
58
+ "Bottleneck - bottleneck.convnext_block.pwconv2": 16448,
59
+ "Decoder - upsamplers.0": 16448,
60
+ "Decoder - upsamplers.1": 16448,
61
+ "Decoder - upsamplers.2": 16448,
62
+ "Decoder - decoders.0.channel_proj": 8256,
63
+ "Decoder - decoders.0.convnext_block.dwconv": 3200,
64
+ "Decoder - decoders.0.convnext_block.norm": 128,
65
+ "Decoder - decoders.0.convnext_block.pwconv1": 16640,
66
+ "Decoder - decoders.0.convnext_block.pwconv2": 16448,
67
+ "Decoder - decoders.1.channel_proj": 8256,
68
+ "Decoder - decoders.1.convnext_block.dwconv": 3200,
69
+ "Decoder - decoders.1.convnext_block.norm": 128,
70
+ "Decoder - decoders.1.convnext_block.pwconv1": 16640,
71
+ "Decoder - decoders.1.convnext_block.pwconv2": 16448,
72
+ "Decoder - decoders.2.channel_proj": 3104,
73
+ "Decoder - decoders.2.convnext_block.dwconv": 1600,
74
+ "Decoder - decoders.2.convnext_block.norm": 64,
75
+ "Decoder - decoders.2.convnext_block.pwconv1": 4224,
76
+ "Decoder - decoders.2.convnext_block.pwconv2": 4128,
77
+ "Other - final_conv": 99
78
+ },
79
+ "latent_dimensions": {
80
+ "Level_0": {
81
+ "channels": 32,
82
+ "height": 96,
83
+ "width": 96,
84
+ "total_elements": 294912
85
+ },
86
+ "Level_1": {
87
+ "channels": 64,
88
+ "height": 48,
89
+ "width": 48,
90
+ "total_elements": 147456
91
+ },
92
+ "Level_2": {
93
+ "channels": 64,
94
+ "height": 24,
95
+ "width": 24,
96
+ "total_elements": 36864
97
+ },
98
+ "Level_3": {
99
+ "channels": 64,
100
+ "height": 12,
101
+ "width": 12,
102
+ "total_elements": 9216
103
+ }
104
+ },
105
+ "skip_dimensions": {
106
+ "Skip_0": {
107
+ "channels": 32,
108
+ "height": 96,
109
+ "width": 96,
110
+ "total_elements": 294912
111
+ },
112
+ "Skip_1": {
113
+ "channels": 64,
114
+ "height": 48,
115
+ "width": 48,
116
+ "total_elements": 147456
117
+ },
118
+ "Skip_2": {
119
+ "channels": 64,
120
+ "height": 24,
121
+ "width": 24,
122
+ "total_elements": 36864
123
+ }
124
+ },
125
+ "memory_analysis": {
126
+ "peak_memory_inference_mb": 2.14453125,
127
+ "peak_memory_training_mb": 4.2890625,
128
+ "peak_elements": 562176
129
+ }
130
+ },
131
+ "inference_results": {
132
+ "success": true,
133
+ "total_time": 2.742616229923442,
134
+ "inference_time": 0.1458,
135
+ "stdout": "[setupvars.sh] OpenVINO environment initialized\nStarting Myriad inference...\nInput: /home/mount/experiments/exp_1762449133_5525_s96_f32_d3_m1_2_2_2/onnx/sample_input.npy\nModel XML: /home/mount/experiments/exp_1762449133_5525_s96_f32_d3_m1_2_2_2/openvino/model.xml\nModel BIN: /home/mount/experiments/exp_1762449133_5525_s96_f32_d3_m1_2_2_2/openvino/model.bin\n[OK] OpenVINO inference engine imported successfully\nDevice: MYRIAD\nModel XML: /home/mount/experiments/exp_1762449133_5525_s96_f32_d3_m1_2_2_2/openvino/model.xml\nModel BIN: /home/mount/experiments/exp_1762449133_5525_s96_f32_d3_m1_2_2_2/openvino/model.bin\nInput file: /home/mount/experiments/exp_1762449133_5525_s96_f32_d3_m1_2_2_2/onnx/sample_input.npy\nInitializing OpenVINO Runtime Core...\nAvailable devices:\n[E:] [BSL] found 0 ioexpander device\n ['CPU', 'GNA', 'MYRIAD']\nLoading network...\nInput blob: input\nInput shape: [1, 8, 96, 96]\nOutput blob: Conv_105\nOutput shape: [1, 3, 96, 96]\nLoading network to MYRIAD...\nLoading input data...\nInput data shape: (1, 8, 96, 96)\nInput data type: float32\nRunning inference...\n[OK] Inference completed!\nInference time: 0.145800 seconds\nOutput shape: (1, 3, 96, 96)\nOutput dtype: float32\nOutput range: [-0.636230, 0.301270]\nMyriad inference completed!\n",
136
+ "stderr": ""
137
+ },
138
+ "end_time": "2025-11-06T18:12:20.507575"
139
+ }