Python: Bicubic interpolation using the GPU … here is a solution to the problem.
Python: Bicubic interpolation using the GPU
In python, I used scipy.ndimage.zoom
to perform bicubic interpolation on images, but found that upsampling was too slow. I would like to replace it with some other GPU-enabled python library, but unfortunately I can’t find one available for python.
NVIDIA provides a good example, which happens to be in C/C++ Bicubic interpolation is implemented in . Are they well-known equivalents< a href="https://developer.nvidia.com/how-to-cuda-python" rel="noreferrer noopener nofollow">CUDA Python example/library that we can directly use and replace scipy.ndimage.zoom
?
I did some searching online but couldn’t find bicubic interpolation using GPU in python. So I don’t think there are many answers available, which leads to self-righteous answers and spam.