Debugging common error messages#

Error messages#

INFO:tensorflow:Error reported to Coordinator: Failed to get convolution algorithm#

INFO:tensorflow:global_step/sec: 0
2019-05-17 13:04:53.550306: E tensorflow/stream_executor/cuda/cuda_dnn.cc:334] Could not create cudnn handle: CUDNN_STATUS_INTERNAL_ERROR
2019-05-17 13:04:53.565467: E tensorflow/stream_executor/cuda/cuda_dnn.cc:334] Could not create cudnn handle: CUDNN_STATUS_INTERNAL_ERROR
INFO:tensorflow:Error reported to Coordinator: Failed to get convolution algorithm. This is probably because cuDNN failed to initialize, so try looking to see if a warning log message was printed above.

If you see an error like this, it’s likely to be one of two things, either your GPU memory is full, or an issue with your CUDA and cuDNN version.

Your GPU memory may be full if it is still being used by another process. To test this, run nvidia-smi, and you will see something like this:

+-----------------------------------------------------------------------------+
| NVIDIA-SMI 418.56       Driver Version: 418.56       CUDA Version: 10.1     |
|-------------------------------+----------------------+----------------------+
| GPU  Name        Persistence-M| Bus-Id        Disp.A | Volatile Uncorr. ECC |
| Fan  Temp  Perf  Pwr:Usage/Cap|         Memory-Usage | GPU-Util  Compute M. |
|===============================+======================+======================|
|   0  TITAN RTX           On   | 00000000:2D:00.0  On |                  N/A |
| 41%   38C    P2    55W / 280W |  23408MiB / 24187MiB |      4%      Default |
+-------------------------------+----------------------+----------------------+

The bit to look for is the memory use (23408MiB / 24187MiB), is this is nearly full (like the example) then find the culprit:

+-----------------------------------------------------------------------------+
| Processes:                                                       GPU Memory |
|  GPU       PID   Type   Process name                             Usage      |
|=============================================================================|
|    0     37793      C   ...miniconda3/envs/cellfinder/bin/python 23408MiB   |
+-----------------------------------------------------------------------------+

In this case, a previous run of cellfinder hasn’t completed. Either wait for it to run, or cancel it with CTRL+C (in the terminal).

Alternatively, your version of CUDA and cuDNN may be not compatible with TensorFlow. You can update them by following the instructions here.

ImportError: DLL load failed: The specified module could not be found.#

ImportError: DLL load failed: The specified module could not be found.

If this occurs on Windows, it is likely that you need to install “Microsoft Visual C++ Redistributable for Visual Studio 2015, 2017 and 2019”, available from Microsoft’s website here.

Things that look like errors, but aren’t:#

Most things that are actually errors will interrupt cellfinder, and the program won’t run. Other things will get logged with an ERROR or a WARNING flag and will get printed to the console in addition to the log file.

A number of third party modules may raise their own errors. As long as you understand what they mean, they can usually be safely ignored.

Can’t find openCV#

CRITICAL:tensorflow:Optional Python module cv2 not found, please install cv2 and retry if the application fails.

Tensorflow thinks this is critical, it’s not.

CPU instruction sets#

tensorflow/core/platform/cpu_feature_guard.cc:141] Your CPU supports instructions that this TensorFlow binary 
was not compiled to use: SSE4.1 SSE4.2 AVX AVX2 FMA

Unless you built tensorflow from source, something like this will come up. It’ll still work fine