in my workflow I am computing the limits by throwing toys with
root -l -b -q StandardHypoTestInvDemo.C > /dev/null
. I noticed that when I set the number of toys in this macro to 500, the job runs fine and produces an output file in about two hours. If I however set N toys to 1000, the job runs for three hours and (from the looks of it) gets killed, and the execution jumps to the next line in the
script:. And I see the
OOMKilled in the last line in the log. I guess I’m hitting some sort of limit: time or memory. When I run on HTCondor (regular submission, nothing to do with Docker or
yadage or anything), setting the N toys even to 20000 does not lead to any issues. So how can I understand the reason for REANA killing the job?
Link to the problematic job: https://reana.cern.ch/details/056b3b5e-30f6-47b8-a5dc-e717b6960cb0, check out the log from the last step.