I'd like to know if my Nvidia GPU, CUDA drivers, cuDNN, Pytorch and TensorFlow are all compatible between each other ahead of time instead of getting some less explicit error when running code such as:
tensorflow/compiler/mlir/tools/kernel_gen/tf_gpu_runtime_wrappers.cc:40] 'cuModuleLoadData(&module, data)' failed with 'CUDA_ERROR_UNSUPPORTED_PTX_VERSION'
How can one automatically check if one's Nvidia GPU, CUDA drivers, cuDNN, Pytorch and TensorFlow are all compatible between each other?