Deep Learning GPU CODER
Show older comments
I have question about Deep Learning Training, GPU Coder and Precision
- We are working on a project that trains data on pre-trained models such as Resnet50 and AlexNet. I would like to use GPU CODER because I heard that using Tensor Core can achieve faster speed efficiency. However, when I looked at the examples related to GPU CODER, there were only examples of data validation using the model. I wonder if it is impossible to train data on a pre-trained model using GPU CODER. If possible, I'd like to ask if there are any relevant examples
- I know that MATLAB uses Single Precision by default when Deep Learning Training. How to use Double Precision or Half Precision?
Accepted Answer
More Answers (0)
Categories
Find more on Parallel and Cloud in Help Center and File Exchange
Community Treasure Hunt
Find the treasures in MATLAB Central and discover how the community can help you!
Start Hunting!