[HDL Coder] Ridiculous RAM usage + How to activate parallel processing?

4 views (last 30 days)
On my project, the HDL Coder uses ridiculus amounts of RAM. On my 8GB Windows 10 Notebook, it chews through it until after 3min the graphics crash. Now I got a 132GB CentOS6 Server at hand and 120GB are occupied after 10min. After 30min, the OOM Killer kicks in and kills the MATLAB process, because it uses all memory and SWAP.
  • Is this RAM usage still sane for the HDL Coder?
  • Is there anyway I could limit this RAM usage?
  • Can I somehow activate parallel computing on the HDL Coder? I think on the server, there was only one core used.
I cannot share the code of the project though.
  1 Comment
Jan Siegmund
Jan Siegmund on 3 Apr 2020
Update: I tried to limit matlabs memory usage to 50GB in Linux:
ulimit -v 50000000
But then it crashes with: Exception found: st9bad_alloc

Sign in to comment.

Accepted Answer

Jan Siegmund
Jan Siegmund on 8 Apr 2020
This memory usage is not sane and it is the result of not properly optimized MATLAB code for HDL conversion. I tried to run code, which processes 4K Images as a whole.
I expected the HDL Coder to serialize the Matrix and processing structures, but this has to be done by hand.
Now my Images are processed in little 5x5 windows, and the HDL coder does not even exceed 1GB of RAM during compilation.
  1 Comment
Jan Siegmund
Jan Siegmund on 8 Apr 2020
However I would love a future feature of the HDL coder, which serializes the processing on its own, because only then, own could use the same speedy processing on FPGA and CPU.

Sign in to comment.

More Answers (0)

Tags

Products


Release

R2020a

Community Treasure Hunt

Find the treasures in MATLAB Central and discover how the community can help you!

Start Hunting!