Skip to content

Commit 7a07ab2

Browse files
committedJul 21, 2017
Merge branch 'master' of github.com:epigen/looper
2 parents df87300 + eb83d9e commit 7a07ab2

File tree

3 files changed

+54
-67
lines changed

3 files changed

+54
-67
lines changed
 

‎doc/source/changelog.rst

+5-1
Original file line numberDiff line numberDiff line change
@@ -1,7 +1,7 @@
11
Changelog
22
******************************
33

4-
- **v0.6** (*unreleased*):
4+
- **v0.6** (*2017-07-21*):
55

66
- New
77

@@ -15,6 +15,8 @@ Changelog
1515

1616
- Allow tsv format for sample sheets.
1717

18+
- Checks that the path to a pipeline actually exists before writing the submission script.
19+
1820
- Changed
1921

2022
- Changed LOOPERENV environment variable to PEPENV, generalizing it to generic models
@@ -29,6 +31,8 @@ Changelog
2931

3032
- Require `setuptools` for installation, and `pandas 0.20.2`. If `numexpr` is installed, version `2.6.2` is required.
3133

34+
- Allows tilde in ``pipeline_interfaces``
35+
3236
- **v0.5** (*2017-03-01*):
3337

3438
- New

‎looper/looper.py

+32-55
Original file line numberDiff line numberDiff line change
@@ -42,7 +42,8 @@ def parse_arguments():
4242
Argument Parsing.
4343
4444
:return argparse.Namespace, list[str]: namespace parsed according to
45-
arguments defined here, them undefined arguments
45+
arguments defined here, and additional options arguments undefined
46+
here and to be handled downstream
4647
"""
4748

4849
# Main looper program help text messages
@@ -390,26 +391,37 @@ def run(prj, args, remaining_args):
390391
_LOGGER.warn("Submission settings "
391392
"lack memory specification")
392393

393-
# Add the command string and job name to the submit_settings object
394+
# Add command string and job name to the submit_settings object.
394395
submit_settings["JOBNAME"] = \
395396
sample.sample_name + "_" + pipeline_key
396397
submit_settings["CODE"] = cmd
397398

398-
# Submit job!
399-
_LOGGER.debug("Attempting job submission: '%s' ('%s')",
400-
sample.sample_name, pl_name)
401-
submitted = cluster_submit(
402-
sample, prj.compute.submission_template,
403-
prj.compute.submission_command, submit_settings,
404-
prj.metadata.submission_subdir, sample_output_folder,
405-
pl_name, args.time_delay, submit=True,
406-
dry_run=args.dry_run, ignore_flags=args.ignore_flags,
407-
remaining_args=remaining_args)
408-
if submitted:
399+
# Create submission script (write script to disk)!
400+
_LOGGER.debug("Creating submission script for pipeline %s: '%s'",
401+
pl_name, sample.sample_name)
402+
submit_script = create_submission_script(
403+
sample, prj.compute.submission_template, submit_settings,
404+
submission_folder=prj.metadata.submission_subdir,
405+
pipeline_name=pl_name, remaining_args=remaining_args)
406+
407+
# Determine how to update submission counts and (perhaps) submit.
408+
flag_files = glob.glob(os.path.join(
409+
sample_output_folder, pl_name + "*.flag"))
410+
if not args.ignore_flags and len(flag_files) > 0:
411+
_LOGGER.info("> Not submitting, flag(s) found: {}".
412+
format(flag_files))
413+
_LOGGER.debug("NOT SUBMITTED")
414+
else:
415+
if args.dry_run:
416+
_LOGGER.info("> DRY RUN: I would have submitted this: '%s'",
417+
submit_script)
418+
else:
419+
submission_command = "{} {}".format(
420+
prj.compute.submission_command, submit_script)
421+
subprocess.call(submission_command, shell=True)
422+
time.sleep(args.time_delay) # Delay next job's submission.
409423
_LOGGER.debug("SUBMITTED")
410424
submit_count += 1
411-
else:
412-
_LOGGER.debug("NOT SUBMITTED")
413425

414426
# Report what went down.
415427
_LOGGER.info("Looper finished")
@@ -630,38 +642,22 @@ def _submission_status_text(curr, total, sample_name, sample_library):
630642

631643

632644

633-
def cluster_submit(
634-
sample, submit_template, submission_command, variables_dict,
635-
submission_folder, sample_output_folder, pipeline_name, time_delay,
636-
submit=False, dry_run=False, ignore_flags=False, remaining_args=None):
645+
def create_submission_script(
646+
sample, submit_template, variables_dict,
647+
submission_folder, pipeline_name, remaining_args=None):
637648
"""
638649
Write cluster submission script to disk and submit job for given Sample.
639650
640651
:param models.Sample sample: the Sample object for submission
641652
:param str submit_template: path to submission script template
642-
:param str submission_command: actual command with which to execute the
643-
submission of the cluster job for the given sample
644653
:param variables_dict: key-value pairs to use to populate fields in
645654
the submission template
646655
:param str submission_folder: path to the folder in which to place
647656
submission files
648-
:param str sample_output_folder: path to folder into which the pipeline
649-
will write file(s), and where to search for flag file to check
650-
if a sample's already been submitted
651657
:param str pipeline_name: name of the pipeline that the job will run
652-
:param int time_delay: number of seconds by which to delay submission
653-
of next job
654-
:param bool submit: whether to even attempt to actually submit the job;
655-
this is useful for skipping certain samples within a project
656-
:param bool dry_run: whether the call is a test and thus the cluster job
657-
created should not actually be submitted; in this case, the return
658-
is a true proxy for whether the job would've been submitted
659-
:param bool ignore_flags: whether to ignore the presence of flag file(s)
660-
in making the determination of whether to submit the job
661658
:param Iterable[str] remaining_args: arguments for this submission,
662659
unconsumed by previous option/argument parsing
663-
:return bool: whether the submission was done,
664-
or would've been if not a dry run
660+
:return str: filepath to submission script
665661
"""
666662

667663
# Create the script and logfile paths.
@@ -706,26 +702,7 @@ def cluster_submit(
706702
name_sample_subtype, sample.name)
707703
sample.to_yaml(subs_folder_path=submission_folder)
708704

709-
# Check if job is already submitted (unless ignore_flags is set to True)
710-
if not ignore_flags:
711-
flag_files = glob.glob(os.path.join(
712-
sample_output_folder, pipeline_name + "*.flag"))
713-
if len(flag_files) > 0:
714-
_LOGGER.info("> Not submitting, flag(s) found: {}".
715-
format(flag_files))
716-
submit = False
717-
else:
718-
pass
719-
720-
if not submit:
721-
return False
722-
if dry_run:
723-
_LOGGER.info("> DRY RUN: I would have submitted this: '%s'",
724-
submit_script)
725-
else:
726-
subprocess.call(submission_command + " " + submit_script, shell=True)
727-
time.sleep(time_delay) # Delay next job's submission.
728-
return True
705+
return submit_script
729706

730707

731708

‎looper/models.py

+17-11
Original file line numberDiff line numberDiff line change
@@ -925,6 +925,13 @@ def build_submission_bundles(self, protocol, priority=True):
925925
strict_pipe_key, full_pipe_path, full_pipe_path_with_flags = \
926926
proto_iface.finalize_pipeline_key_and_paths(
927927
pipeline_key)
928+
929+
# Skip and warn about nonexistent alleged pipeline path.
930+
if not _os.path.exists(full_pipe_path):
931+
_LOGGER.warn(
932+
"Missing pipeline script: '%s'", full_pipe_path)
933+
continue
934+
928935
# Determine which interface and Sample subtype to use.
929936
sample_subtype = \
930937
proto_iface.fetch_sample_subtype(
@@ -1128,7 +1135,7 @@ def parse_config_file(self, subproject=None):
11281135
# Parse yaml into the project's attributes.
11291136
_LOGGER.debug("Adding attributes for {}: {}".format(
11301137
self.__class__.__name__, config.keys()))
1131-
_LOGGER.debug("Config metadata: {}")
1138+
_LOGGER.debug("Config metadata: {}".format(config["metadata"]))
11321139
self.add_entries(config)
11331140
_LOGGER.debug("{} now has {} keys: {}".format(
11341141
self.__class__.__name__, len(self.keys()), self.keys()))
@@ -1348,7 +1355,7 @@ def _ensure_absolute(self, maybe_relpath):
13481355
_LOGGER.log(5, "Already absolute")
13491356
return maybe_relpath
13501357
# Maybe we have env vars that make the path absolute?
1351-
expanded = _os.path.expandvars(maybe_relpath)
1358+
expanded = _os.path.expanduser(_os.path.expandvars(maybe_relpath))
13521359
_LOGGER.log(5, "Expanded: '%s'", expanded)
13531360
if _os.path.isabs(expanded):
13541361
_LOGGER.log(5, "Expanded is absolute")
@@ -1834,8 +1841,7 @@ def set_transcriptome(self, transcriptomes):
18341841

18351842
def _set_assembly(self, ome, assemblies):
18361843
if not assemblies:
1837-
_LOGGER.debug("Empty/null assemblies mapping: {} ({})".
1838-
format(assemblies, type(assemblies)))
1844+
_LOGGER.debug("Empty/null assemblies mapping")
18391845
return
18401846
try:
18411847
assembly = assemblies[self.organism]
@@ -2619,11 +2625,14 @@ def finalize_pipeline_key_and_paths(self, pipeline_key):
26192625
# The strict key is the script name itself, something like "ATACseq.py"
26202626
strict_pipeline_key, _, pipeline_key_args = pipeline_key.partition(' ')
26212627

2622-
if self.pipe_iface.get_attribute(strict_pipeline_key, "path"):
2623-
script_path_only = self.pipe_iface.get_attribute(
2624-
strict_pipeline_key, "path")[0].strip()
2628+
full_pipe_path = \
2629+
self.pipe_iface.get_attribute(strict_pipeline_key, "path")
2630+
if full_pipe_path:
2631+
script_path_only = _os.path.expanduser(_os.path.expandvars(full_pipe_path[0].strip()))
2632+
if _os.path.isdir(script_path_only):
2633+
script_path_only = _os.path.join(script_path_only, pipeline_key)
26252634
script_path_with_flags = \
2626-
" ".join([script_path_only, pipeline_key_args])
2635+
"{} {}".format(script_path_only, pipeline_key_args)
26272636
else:
26282637
# backwards compatibility w/ v0.5
26292638
script_path_only = strict_pipeline_key
@@ -2639,9 +2648,6 @@ def finalize_pipeline_key_and_paths(self, pipeline_key):
26392648
self.pipelines_path, script_path_with_flags)
26402649
_LOGGER.log(5, "Absolute script path with flags: '%s'",
26412650
script_path_with_flags)
2642-
if not _os.path.exists(script_path_only):
2643-
_LOGGER.warn(
2644-
"Missing pipeline script: '%s'", script_path_only)
26452651

26462652
return strict_pipeline_key, script_path_only, script_path_with_flags
26472653

0 commit comments

Comments
 (0)
Please sign in to comment.