Dear yc.lee,
Below are the logs for your three failed submissions.
test13
2022-02-15T13:41:53+00:00
0%| | 0/234 [00:00<?, ?it/s]2022-02-15 22:41:53.228917: I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:939] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero
2022-02-15T13:41:53+00:00 2022-02-15 22:41:53.237291: I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:939] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero
2022-02-15T13:41:53+00:00 2022-02-15 22:41:53.237940: I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:939] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero
2022-02-15T13:41:53+00:00 2022-02-15 22:41:53.238788: I tensorflow/core/platform/cpu_feature_guard.cc:151] This TensorFlow binary is optimized with oneAPI Deep Neural Network Library (oneDNN) to use the following CPU instructions in performance-critical operations: AVX2 AVX512F FMA
2022-02-15T13:41:53+00:00 To enable them in other operations, rebuild TensorFlow with the appropriate compiler flags.
2022-02-15T13:41:53+00:00 2022-02-15 22:41:53.239443: I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:939] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero
2022-02-15T13:41:53+00:00 2022-02-15 22:41:53.240081: I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:939] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero
2022-02-15T13:41:53+00:00 2022-02-15 22:41:53.240661: I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:939] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero
2022-02-15T13:41:54+00:00 2022-02-15 22:41:54.082353: I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:939] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero
2022-02-15T13:41:54+00:00 2022-02-15 22:41:54.083026: I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:939] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero
2022-02-15T13:41:54+00:00 2022-02-15 22:41:54.083628: I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:939] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero
2022-02-15T13:41:54+00:00 2022-02-15 22:41:54.084201: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1525] Created device /job:localhost/replica:0/task:0/device:GPU:0 with 13793 MB memory: -> device: 0, name: Tesla T4, pci bus id: 0000:00:1e.0, compute capability: 7.5
2022-02-15T13:42:00+00:00 2022-02-15 22:42:00.969723: I tensorflow/stream_executor/cuda/cuda_dnn.cc:366] Loaded cuDNN version 8204
2022-02-15T13:42:01+00:00 2022-02-15 22:42:01.631010: I tensorflow/core/platform/default/subprocess.cc:304] Start cannot spawn child process: No such file or directory
2022-02-15T13:42:01+00:00 2022-02-15 22:42:01.631453: I tensorflow/core/platform/default/subprocess.cc:304] Start cannot spawn child process: No such file or directory
2022-02-15T13:42:01+00:00 2022-02-15 22:42:01.631473: W tensorflow/stream_executor/gpu/asm_compiler.cc:80] Couldn't get ptxas version string: INTERNAL: Couldn't invoke ptxas --version
2022-02-15T13:42:01+00:00 2022-02-15 22:42:01.631918: I tensorflow/core/platform/default/subprocess.cc:304] Start cannot spawn child process: No such file or directory
2022-02-15T13:42:01+00:00 2022-02-15 22:42:01.631986: W tensorflow/stream_executor/gpu/redzone_allocator.cc:314] INTERNAL: Failed to launch ptxas
2022-02-15T13:42:01+00:00 Relying on driver to perform ptx compilation.
2022-02-15T13:42:01+00:00 Modify $PATH to customize ptxas location.
2022-02-15T13:42:01+00:00 This message will be only logged once.
2022-02-15T13:42:06+00:00 WARNING:tensorflow:Unresolved object in checkpoint: (root).optimizer
2022-02-15T13:42:06+00:00 WARNING:tensorflow:Unresolved object in checkpoint: (root).optimizer.decay
2022-02-15T13:42:06+00:00 WARNING:tensorflow:Unresolved object in checkpoint: (root).optimizer.learning_rate
2022-02-15T13:42:06+00:00 WARNING:tensorflow:Unresolved object in checkpoint: (root).optimizer.momentum
2022-02-15T13:42:06+00:00 WARNING:tensorflow:Unresolved object in checkpoint: (root).optimizer.iter
2022-02-15T13:42:06+00:00 WARNING:tensorflow:A checkpoint was restored (e.g. tf.train.Checkpoint.restore or tf.keras.Model.load_weights) but not all checkpointed values were used. See above for specific issues. Use expect_partial() on the load status object, e.g. tf.train.Checkpoint.restore(...).expect_partial(), to silence these warnings, or use assert_consumed() to make the check explicit. See https://www.tensorflow.org/guide/checkpoint#loading_mechanics for details.
2022-02-15T13:42:11+00:00
0%| | 0/234 [00:21<?, ?it/s]
2022-02-15T13:42:11+00:00 Traceback (most recent call last):
2022-02-15T13:42:11+00:00 File "/usr/lib/python3.8/runpy.py", line 194, in _run_module_as_main
2022-02-15T13:42:11+00:00 return _run_code(code, main_globals, None,
2022-02-15T13:42:11+00:00 File "/usr/lib/python3.8/runpy.py", line 87, in _run_code
2022-02-15T13:42:11+00:00 exec(code, run_globals)
2022-02-15T13:42:11+00:00 File "/opt/algorithm/process.py", line 180, in <module>
2022-02-15T13:42:11+00:00 airogs_algorithm().process()
2022-02-15T13:42:11+00:00 File "/home/algorithm/.local/lib/python3.8/site-packages/evalutils/evalutils.py", line 183, in process
2022-02-15T13:42:11+00:00 self.process_cases()
2022-02-15T13:42:11+00:00 File "/home/algorithm/.local/lib/python3.8/site-packages/evalutils/evalutils.py", line 191, in process_cases
2022-02-15T13:42:11+00:00 self._case_results.append(self.process_case(idx=idx, case=case))
2022-02-15T13:42:11+00:00 File "/opt/algorithm/process.py", line 93, in process_case
2022-02-15T13:42:11+00:00 results.append(self.predict(input_image_array=input_image_array))
2022-02-15T13:42:11+00:00 File "/opt/algorithm/process.py", line 134, in predict
2022-02-15T13:42:11+00:00 pred_y = model([disc_img, fundus_img], training=True)
2022-02-15T13:42:11+00:00 File "/home/algorithm/.local/lib/python3.8/site-packages/keras/utils/traceback_utils.py", line 67, in error_handler
2022-02-15T13:42:11+00:00 raise e.with_traceback(filtered_tb) from None
2022-02-15T13:42:11+00:00 File "/home/algorithm/.local/lib/python3.8/site-packages/keras/engine/input_spec.py", line 263, in assert_input_compatibility
2022-02-15T13:42:11+00:00 raise ValueError(f'Input {input_index} of layer "{layer_name}" is '
2022-02-15T13:42:11+00:00 ValueError: Input 0 of layer "model_1" is incompatible with the layer: expected shape=(None, 608, 608, 1), found shape=(1, 608, 559, 1)
test14
2022-02-16T04:08:41+00:00 2022-02-16 13:08:41.620931: I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:939] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero
2022-02-16T04:08:41+00:00 2022-02-16 13:08:41.630304: I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:939] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero
2022-02-16T04:08:41+00:00 2022-02-16 13:08:41.630980: I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:939] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero
2022-02-16T04:08:41+00:00 2022-02-16 13:08:41.631787: I tensorflow/core/platform/cpu_feature_guard.cc:151] This TensorFlow binary is optimized with oneAPI Deep Neural Network Library (oneDNN) to use the following CPU instructions in performance-critical operations: AVX2 AVX512F FMA
2022-02-16T04:08:41+00:00 To enable them in other operations, rebuild TensorFlow with the appropriate compiler flags.
2022-02-16T04:08:41+00:00 2022-02-16 13:08:41.632459: I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:939] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero
2022-02-16T04:08:41+00:00 2022-02-16 13:08:41.633066: I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:939] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero
2022-02-16T04:08:41+00:00 2022-02-16 13:08:41.633638: I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:939] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero
2022-02-16T04:08:42+00:00 2022-02-16 13:08:42.223436: I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:939] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero
2022-02-16T04:08:42+00:00 2022-02-16 13:08:42.224148: I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:939] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero
2022-02-16T04:08:42+00:00 2022-02-16 13:08:42.224758: I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:939] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero
2022-02-16T04:08:42+00:00 2022-02-16 13:08:42.225326: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1525] Created device /job:localhost/replica:0/task:0/device:GPU:0 with 13793 MB memory: -> device: 0, name: Tesla T4, pci bus id: 0000:00:1e.0, compute capability: 7.5
2022-02-16T04:08:56+00:00
0%| | 0/234 [00:00<?, ?it/s]2022-02-16 13:08:56.002094: I tensorflow/stream_executor/cuda/cuda_dnn.cc:366] Loaded cuDNN version 8204
2022-02-16T04:08:56+00:00 2022-02-16 13:08:56.674469: I tensorflow/core/platform/default/subprocess.cc:304] Start cannot spawn child process: No such file or directory
2022-02-16T04:08:56+00:00 2022-02-16 13:08:56.674960: I tensorflow/core/platform/default/subprocess.cc:304] Start cannot spawn child process: No such file or directory
2022-02-16T04:08:56+00:00 2022-02-16 13:08:56.674984: W tensorflow/stream_executor/gpu/asm_compiler.cc:80] Couldn't get ptxas version string: INTERNAL: Couldn't invoke ptxas --version
2022-02-16T04:08:56+00:00 2022-02-16 13:08:56.675409: I tensorflow/core/platform/default/subprocess.cc:304] Start cannot spawn child process: No such file or directory
2022-02-16T04:08:56+00:00 2022-02-16 13:08:56.675471: W tensorflow/stream_executor/gpu/redzone_allocator.cc:314] INTERNAL: Failed to launch ptxas
2022-02-16T04:08:56+00:00 Relying on driver to perform ptx compilation.
2022-02-16T04:08:56+00:00 Modify $PATH to customize ptxas location.
2022-02-16T04:08:56+00:00 This message will be only logged once.
2022-02-16T04:09:01+00:00
0%| | 0/234 [00:09<?, ?it/s]
2022-02-16T04:09:01+00:00 Traceback (most recent call last):
2022-02-16T04:09:01+00:00 File "/usr/lib/python3.8/runpy.py", line 194, in _run_module_as_main
2022-02-16T04:09:01+00:00 return _run_code(code, main_globals, None,
2022-02-16T04:09:01+00:00 File "/usr/lib/python3.8/runpy.py", line 87, in _run_code
2022-02-16T04:09:01+00:00 exec(code, run_globals)
2022-02-16T04:09:01+00:00 File "/opt/algorithm/process.py", line 185, in <module>
2022-02-16T04:09:01+00:00 airogs_algorithm().process()
2022-02-16T04:09:01+00:00 File "/home/algorithm/.local/lib/python3.8/site-packages/evalutils/evalutils.py", line 183, in process
2022-02-16T04:09:01+00:00 self.process_cases()
2022-02-16T04:09:01+00:00 File "/home/algorithm/.local/lib/python3.8/site-packages/evalutils/evalutils.py", line 191, in process_cases
2022-02-16T04:09:01+00:00 self._case_results.append(self.process_case(idx=idx, case=case))
2022-02-16T04:09:01+00:00 File "/opt/algorithm/process.py", line 107, in process_case
2022-02-16T04:09:01+00:00 results.append(self.predict(cls_model=cls_model, seg_model=seg_model, input_image_array=input_image_array))
2022-02-16T04:09:01+00:00 File "/opt/algorithm/process.py", line 141, in predict
2022-02-16T04:09:01+00:00 pred_y = cls_model([disc_img, fundus_img], training=True)
2022-02-16T04:09:01+00:00 File "/home/algorithm/.local/lib/python3.8/site-packages/keras/utils/traceback_utils.py", line 67, in error_handler
2022-02-16T04:09:01+00:00 raise e.with_traceback(filtered_tb) from None
2022-02-16T04:09:01+00:00 File "/home/algorithm/.local/lib/python3.8/site-packages/keras/engine/input_spec.py", line 263, in assert_input_compatibility
2022-02-16T04:09:01+00:00 raise ValueError(f'Input {input_index} of layer "{layer_name}" is '
2022-02-16T04:09:01+00:00 ValueError: Input 0 of layer "model_1" is incompatible with the layer: expected shape=(None, 608, 608, 1), found shape=(1, 608, 559, 1)
2022-02-16T04:09:02+00:00 WARNING:tensorflow:Unresolved object in checkpoint: (root).optimizer
2022-02-16T04:09:02+00:00 WARNING:tensorflow:Unresolved object in checkpoint: (root).optimizer.decay
2022-02-16T04:09:02+00:00 WARNING:tensorflow:Unresolved object in checkpoint: (root).optimizer.learning_rate
2022-02-16T04:09:02+00:00 WARNING:tensorflow:Unresolved object in checkpoint: (root).optimizer.momentum
2022-02-16T04:09:02+00:00 WARNING:tensorflow:Unresolved object in checkpoint: (root).optimizer.iter
2022-02-16T04:09:02+00:00 WARNING:tensorflow:A checkpoint was restored (e.g. tf.train.Checkpoint.restore or tf.keras.Model.load_weights) but not all checkpointed values were used. See above for specific issues. Use expect_partial() on the load status object, e.g. tf.train.Checkpoint.restore(...).expect_partial(), to silence these warnings, or use assert_consumed() to make the check explicit. See https://www.tensorflow.org/guide/checkpoint#loading_mechanics for details.
test15
2022-02-16T06:45:24+00:00 2022-02-16 15:45:24.822555: I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:939] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero
2022-02-16T06:45:24+00:00 2022-02-16 15:45:24.832034: I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:939] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero
2022-02-16T06:45:24+00:00 2022-02-16 15:45:24.832676: I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:939] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero
2022-02-16T06:45:24+00:00 2022-02-16 15:45:24.833496: I tensorflow/core/platform/cpu_feature_guard.cc:151] This TensorFlow binary is optimized with oneAPI Deep Neural Network Library (oneDNN) to use the following CPU instructions in performance-critical operations: AVX2 AVX512F FMA
2022-02-16T06:45:24+00:00 To enable them in other operations, rebuild TensorFlow with the appropriate compiler flags.
2022-02-16T06:45:24+00:00 2022-02-16 15:45:24.834132: I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:939] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero
2022-02-16T06:45:24+00:00 2022-02-16 15:45:24.834740: I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:939] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero
2022-02-16T06:45:24+00:00 2022-02-16 15:45:24.835338: I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:939] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero
2022-02-16T06:45:25+00:00 2022-02-16 15:45:25.764199: I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:939] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero
2022-02-16T06:45:25+00:00 2022-02-16 15:45:25.764899: I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:939] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero
2022-02-16T06:45:25+00:00 2022-02-16 15:45:25.765494: I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:939] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero
2022-02-16T06:45:25+00:00 2022-02-16 15:45:25.766054: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1525] Created device /job:localhost/replica:0/task:0/device:GPU:0 with 13793 MB memory: -> device: 0, name: Tesla T4, pci bus id: 0000:00:1e.0, compute capability: 7.5
2022-02-16T06:45:40+00:00
0%| | 0/234 [00:00<?, ?it/s]2022-02-16 15:45:40.081145: I tensorflow/stream_executor/cuda/cuda_dnn.cc:366] Loaded cuDNN version 8204
2022-02-16T06:45:40+00:00 2022-02-16 15:45:40.767710: I tensorflow/core/platform/default/subprocess.cc:304] Start cannot spawn child process: No such file or directory
2022-02-16T06:45:40+00:00 2022-02-16 15:45:40.768169: I tensorflow/core/platform/default/subprocess.cc:304] Start cannot spawn child process: No such file or directory
2022-02-16T06:45:40+00:00 2022-02-16 15:45:40.768192: W tensorflow/stream_executor/gpu/asm_compiler.cc:80] Couldn't get ptxas version string: INTERNAL: Couldn't invoke ptxas --version
2022-02-16T06:45:40+00:00 2022-02-16 15:45:40.768679: I tensorflow/core/platform/default/subprocess.cc:304] Start cannot spawn child process: No such file or directory
2022-02-16T06:45:40+00:00 2022-02-16 15:45:40.768753: W tensorflow/stream_executor/gpu/redzone_allocator.cc:314] INTERNAL: Failed to launch ptxas
2022-02-16T06:45:40+00:00 Relying on driver to perform ptx compilation.
2022-02-16T06:45:40+00:00 Modify $PATH to customize ptxas location.
2022-02-16T06:45:40+00:00 This message will be only logged once.
2022-02-16T06:45:45+00:00
0%| | 0/234 [00:09<?, ?it/s]
2022-02-16T06:45:45+00:00 Traceback (most recent call last):
2022-02-16T06:45:45+00:00 File "/usr/lib/python3.8/runpy.py", line 194, in _run_module_as_main
2022-02-16T06:45:45+00:00 return _run_code(code, main_globals, None,
2022-02-16T06:45:45+00:00 File "/usr/lib/python3.8/runpy.py", line 87, in _run_code
2022-02-16T06:45:45+00:00 exec(code, run_globals)
2022-02-16T06:45:45+00:00 File "/opt/algorithm/process.py", line 185, in <module>
2022-02-16T06:45:45+00:00 airogs_algorithm().process()
2022-02-16T06:45:45+00:00 File "/home/algorithm/.local/lib/python3.8/site-packages/evalutils/evalutils.py", line 183, in process
2022-02-16T06:45:45+00:00 self.process_cases()
2022-02-16T06:45:45+00:00 File "/home/algorithm/.local/lib/python3.8/site-packages/evalutils/evalutils.py", line 191, in process_cases
2022-02-16T06:45:45+00:00 self._case_results.append(self.process_case(idx=idx, case=case))
2022-02-16T06:45:45+00:00 File "/opt/algorithm/process.py", line 107, in process_case
2022-02-16T06:45:45+00:00 results.append(self.predict(cls_model=cls_model, seg_model=seg_model, input_image_array=input_image_array))
2022-02-16T06:45:45+00:00 File "/opt/algorithm/process.py", line 141, in predict
2022-02-16T06:45:45+00:00 pred_y = cls_model([disc_img, fundus_img], training=True)
2022-02-16T06:45:45+00:00 File "/home/algorithm/.local/lib/python3.8/site-packages/keras/utils/traceback_utils.py", line 67, in error_handler
2022-02-16T06:45:45+00:00 raise e.with_traceback(filtered_tb) from None
2022-02-16T06:45:45+00:00 File "/home/algorithm/.local/lib/python3.8/site-packages/keras/engine/input_spec.py", line 263, in assert_input_compatibility
2022-02-16T06:45:45+00:00 raise ValueError(f'Input {input_index} of layer "{layer_name}" is '
2022-02-16T06:45:45+00:00 ValueError: Input 0 of layer "model_1" is incompatible with the layer: expected shape=(None, 608, 608, 1), found shape=(1, 608, 559, 1)
2022-02-16T06:45:46+00:00 WARNING:tensorflow:Unresolved object in checkpoint: (root).optimizer
2022-02-16T06:45:46+00:00 WARNING:tensorflow:Unresolved object in checkpoint: (root).optimizer.decay
2022-02-16T06:45:46+00:00 WARNING:tensorflow:Unresolved object in checkpoint: (root).optimizer.learning_rate
2022-02-16T06:45:46+00:00 WARNING:tensorflow:Unresolved object in checkpoint: (root).optimizer.momentum
2022-02-16T06:45:46+00:00 WARNING:tensorflow:Unresolved object in checkpoint: (root).optimizer.iter
2022-02-16T06:45:46+00:00 WARNING:tensorflow:A checkpoint was restored (e.g. tf.train.Checkpoint.restore or tf.keras.Model.load_weights) but not all checkpointed values were used. See above for specific issues. Use expect_partial() on the load status object, e.g. tf.train.Checkpoint.restore(...).expect_partial(), to silence these warnings, or use assert_consumed() to make the check explicit. See https://www.tensorflow.org/guide/checkpoint#loading_mechanics for details.
Best regards,
Coen de Vente