Does anyone know where I should start looking to perform matrix-based calculations that are GPU accelerated? I've lately gotten into neural networks, which require a huge amount of matrix math. I've been able to train neural networks on servers/computers, and there are cases where I want to run the network on iOS devices.
I've successfully ported some neural networks, like for handwriting recognition, to Swift.
However, it quickly slows down when I increase the complexity. For example, it will analyze a 20x20 image just fine, but when I start increasing the resolution to 56x56 or more, the number of calculations increases exponentially. It is a perfect scenario for a GPU.
Are there any good matrix libraries/frameworks that let you create variable sized matrices and perform calculations with them, that are GPU accelerated?
Thanks!