How to use GPU to define a large 3D matrix
Show older comments
Dear all,
I am trying to use the attached M file (mkdelta.m) to difine a large 3d matrix (5598x40x40). It will cost large amount of CPU time to finish, which is unbearable. I am considering to use GPU to accelerate the computation. I have GPU in my computer but I have no idea how to use it to do this. Could you give me some tutorial?
Best regards.
Yeping Sun
Accepted Answer
More Answers (0)
Categories
Find more on Get Started with GPU Coder in Help Center and File Exchange
Community Treasure Hunt
Find the treasures in MATLAB Central and discover how the community can help you!
Start Hunting!