<?xml version="1.0" encoding="UTF-8"?>
<feed xmlns="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <title>Optimal Job Size for the SLURM Scheduler?</title>
  <link rel="alternate" href="https://conferences.xsede.org/c/message_boards/find_thread?p_l_id=&amp;threadId=681095" />
  <subtitle>Optimal Job Size for the SLURM Scheduler?</subtitle>
  <entry>
    <title>Optimal Job Size for the SLURM Scheduler?</title>
    <link rel="alternate" href="https://conferences.xsede.org/c/message_boards/find_message?p_l_id=&amp;messageId=681094" />
    <author>
      <name>Tim Supinie</name>
    </author>
    <id>https://conferences.xsede.org/c/message_boards/find_message?p_l_id=&amp;messageId=681094</id>
    <updated>2014-03-21T15:12:38Z</updated>
    <published>2014-03-21T15:12:38Z</published>
    <summary type="html">I developed my experiment setup on Kraken and ported to Stampede a month and a half ago, and for the first month it was great.  In the last week or so, however, my queue throughput has tanked pretty severely.  On Kraken, I think the scheduler favored large jobs, except if you could make them small enough to run as &amp;#034;backfill,&amp;#034; and I managed to split my experiments up into small enough chunks that it was frequently running as backfill.  This brings me to my question: what is the optimal job size?  Is it set up the same as Kraken to where it favors larger jobs, but will backfill?  If so, how small do jobs need to be to run as backfill?  Waiting 4 hours for 32-core-for-20-minute jobs to go through is getting a little old.&lt;br /&gt;&lt;br /&gt;Thanks!</summary>
    <dc:creator>Tim Supinie</dc:creator>
    <dc:date>2014-03-21T15:12:38Z</dc:date>
  </entry>
</feed>

