Skip to content

Commit 3ac4bd0

Browse files
committed
more cleanup
1 parent c4a1831 commit 3ac4bd0

3 files changed

Lines changed: 24 additions & 836 deletions

File tree

_episodes/07-grid-job-submission.md

Lines changed: 24 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -368,6 +368,29 @@ Since the workflow was causing a systemwide disruption we immediately held all o
368368

369369
DUNE has also created a a global glideinWMS pool similar to the CMS Global Pool that is intended to serve as a single point through which multiple job submission systems (e.g. HTCondor schedulers at sites outside of Fermilab) can have access to the same resources. Jobs using the global pool still run in the exactly the same way as those that don't. We plan to move more and more work over to the global pool in 2023 and priority access to the FermiGrid quota will eventually be given to jobs submitted to the global pool. To switch to the global pool with jobsub, it's simply a matter of adding `--global-pool dune` as an option to your submission command. The only practical difference is that your jobs will come back with IDs of the form NNNNNNN.N@dunegpschedd0X.fnal.gov instead of NNNNNNN.N@jobsub0X.fnal.gov. Again, everything else is identical, so feel free to test it out.
370370

371+
372+
## Making subsets of metacat datasets
373+
374+
Running across very large number of files puts you at risk of system issues. It is often much nicer to run over several smaller subsets.
375+
Many official metacat definitions are large data collections defined only by their properties and not really suitable for a single job.
376+
377+
You can do the following. Submit your jobs using the skip and limit commands. Here 'namespace:official_dataset' describes the official dataset.
378+
379+
See [the basics tutorial](https://dune.github.io/computing-basics/03-data-management/index.html#official-datasets-) for information on official datasets.
380+
381+
~~~
382+
query="files from namespace:official_dataset skip 0 limit 1000"
383+
query="files from namespace:official_dataset skip 1000 limit 1000"
384+
query="files from namespace:official_dataset skip 2000 limit 1000"
385+
....
386+
~~~
387+
{: ..language-bash}
388+
389+
390+
391+
392+
<!--
393+
371394
## Making subsets of sam definitions
372395
373396
Running across very large number of files puts you at risk of system issues. It is often much nicer to run over several smaller subsets.
@@ -399,7 +422,7 @@ samweb create-definition $USER-mydataset-part4 “defname:mydataset limit 2000 w
399422
400423
401424
402-
More on samweb can be found [here]({{ site.baseurl }}/sam-by-schellman).
425+
More on samweb can be found [here]({{ site.baseurl }}/sam-by-schellman). -->
403426

404427
## Verify Your Learning:
405428

0 commit comments

Comments
 (0)