You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I wonder if there are either 1) solutions for this or 2) easy ways to add the ability to run a looper pipeline in an ad hoc manner. What I mean by that is this: occasionally, the overhead of a traditional workflow can be a bit daunting, but I really enjoy the ease of dispatching off jobs through slurm+looper.
I would love to replace traditional bash for loops with looper calls.
An example
I have a folder with hundreds of mixed-type files. Some of these might be bedGraph files. I want to convert these to .bw format. I can use bigtools bedGraphToBigWig. Traditionally, I might just use a for loop:
But this takes awhile since it goes one-by-one, and there are hundreds. I'd love to fire them all off at once using looper and slurm:
ls *.bdg | looper run "bigtools bedGraphToBigWig {$1} {$1}.bw"
I suppose I am trying to identify or nail-down a potential gap between traditional workflows and the flexibility researchers often need for quick, ad hoc job submission.
The text was updated successfully, but these errors were encountered:
@nleroy917 this is a good idea. IIRC, way back in time, @nsheff had an example or two like this which sort of "pushed the limits" "/ thought outside the box" (if I'm permitted some clichés) of looper in this way, maybe he has already a working example or something closest to this which would represent a good starting point?
What to do with the command template?
Maybe using -y to give it a command template (command-extra-override) is a way to provide a command template when there was none to begin with?
Can we make a PEP on the fly given some way of info?
Sure... we can make it accept stdin and then what I wrote would work...?
Only works when $(nproc) returns a value greater than one of course... so you still would need to allocate some cores for yourself. Its an interesting stop-gap, but I still think the looper version proposed above would be way better.
I wonder if there are either 1) solutions for this or 2) easy ways to add the ability to run a looper pipeline in an ad hoc manner. What I mean by that is this: occasionally, the overhead of a traditional workflow can be a bit daunting, but I really enjoy the ease of dispatching off jobs through slurm+looper.
I would love to replace traditional bash for loops with
looper
calls.An example
I have a folder with hundreds of mixed-type files. Some of these might be
bedGraph
files. I want to convert these to.bw
format. I can usebigtools bedGraphToBigWig
. Traditionally, I might just use a for loop:But this takes awhile since it goes one-by-one, and there are hundreds. I'd love to fire them all off at once using
looper
and slurm:I suppose I am trying to identify or nail-down a potential gap between traditional workflows and the flexibility researchers often need for quick, ad hoc job submission.
The text was updated successfully, but these errors were encountered: