+ PROC_LSST_SUBMIT_PATH=/mmfs1/home/stevengs/dirac/DEEP/submit + bps submit /gscratch/dirac/shared/opt/proc_lsst/pipelines/submit.yaml -b /mmfs1/home/stevengs/dirac/DEEP/repo -i DEEP/20190601/A1a --output-run DEEP/20190601/A1a/science#step1/20240313T015456Z --qgraph pipeline.qgraph lsst.ctrl.bps.drivers INFO: DISCLAIMER: All values regarding memory consumption reported below are approximate and may not accurately reflect actual memory usage by the bps process. lsst.ctrl.bps.drivers INFO: Starting submission process lsst.ctrl.bps.drivers INFO: Initializing execution environment lsst.ctrl.bps.drivers INFO: Initializing execution environment completed: Took 1.7669 seconds; current memory usage: 0.178 Gibyte, delta: 0.041 Gibyte, peak delta: 0.054 Gibyte lsst.ctrl.bps.drivers INFO: Peak memory usage for bps process 0.190 Gibyte (main), 0.000 Gibyte (largest child process) lsst.ctrl.bps.drivers INFO: Starting acquire stage (generating and/or reading quantum graph) lsst.ctrl.bps.pre_transform INFO: Copying quantum graph from 'pipeline.qgraph' lsst.ctrl.bps.pre_transform INFO: Completed copying quantum graph: Took 0.0024 seconds lsst.ctrl.bps.pre_transform INFO: Backing up quantum graph from '/mmfs1/home/stevengs/dirac/DEEP/submit/DEEP/20190601/A1a/science#step1/20240313T015456Z/submit/DEEP/20190601/A1a/science#step1/20240313T015456Z/pipeline.qgraph' lsst.ctrl.bps.pre_transform INFO: Completed backing up quantum graph: Took 0.0059 seconds lsst.ctrl.bps.pre_transform INFO: /mmfs1/gscratch/dirac/shared/opt/conda/envs/lsst-scipipe-8.0.0/share/eups/Linux64/ctrl_mpexec/g1ce94f1343+c79f27626b/bin/pipetask --long-log --log-level=VERBOSE update-graph-run /mmfs1/home/stevengs/dirac/DEEP/submit/DEEP/20190601/A1a/science#step1/20240313T015456Z/submit/DEEP/20190601/A1a/science#step1/20240313T015456Z/pipeline_orig.qgraph DEEP/20190601/A1a/science#step1/20240313T015456Z /mmfs1/home/stevengs/dirac/DEEP/submit/DEEP/20190601/A1a/science#step1/20240313T015456Z/submit/DEEP/20190601/A1a/science#step1/20240313T015456Z/pipeline.qgraph lsst.ctrl.bps.pre_transform INFO: Reading quantum graph from '/mmfs1/home/stevengs/dirac/DEEP/submit/DEEP/20190601/A1a/science#step1/20240313T015456Z/submit/DEEP/20190601/A1a/science#step1/20240313T015456Z/pipeline.qgraph' lsst.ctrl.bps.pre_transform INFO: Completed reading quantum graph: Took 5.6787 seconds lsst.ctrl.bps.drivers INFO: Acquire stage completed: Took 13.2859 seconds; current memory usage: 0.337 Gibyte, delta: 0.159 Gibyte, peak delta: 0.146 Gibyte lsst.ctrl.bps.drivers INFO: Peak memory usage for bps process 0.337 Gibyte (main), 0.325 Gibyte (largest child process) lsst.ctrl.bps.drivers INFO: Starting cluster stage (grouping quanta into jobs) lsst.ctrl.bps.drivers INFO: Cluster stage completed: Took 0.0116 seconds; current memory usage: 0.337 Gibyte, delta: 0.000 Gibyte, peak delta: 0.000 Gibyte lsst.ctrl.bps.drivers INFO: Peak memory usage for bps process 0.337 Gibyte (main), 0.325 Gibyte (largest child process) lsst.ctrl.bps.drivers INFO: ClusteredQuantumGraph contains 6 cluster(s) lsst.ctrl.bps.drivers INFO: Starting transform stage (creating generic workflow) lsst.ctrl.bps.drivers INFO: Generic workflow name 'DEEP_20190601_A1a_science#step1_20240313T015456Z' lsst.ctrl.bps.drivers INFO: Transform stage completed: Took 0.0293 seconds; current memory usage: 0.337 Gibyte, delta: 0.000 Gibyte, peak delta: 0.000 Gibyte lsst.ctrl.bps.drivers INFO: Peak memory usage for bps process 0.337 Gibyte (main), 0.325 Gibyte (largest child process) lsst.ctrl.bps.drivers INFO: GenericWorkflow contains 8 job(s) (including final) lsst.ctrl.bps.drivers INFO: Starting prepare stage (creating specific implementation of workflow) parsl.addresses ERROR: Ignoring failure to fetch address from interface eno2 Traceback (most recent call last): File "/mmfs1/gscratch/dirac/shared/opt/conda/envs/lsst-scipipe-8.0.0/lib/python3.11/site-packages/parsl/addresses.py", line 111, in get_all_addresses s_addresses.add(address_by_interface(interface)) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/mmfs1/gscratch/dirac/shared/opt/conda/envs/lsst-scipipe-8.0.0/lib/python3.11/site-packages/typeguard/__init__.py", line 1033, in wrapper retval = func(*args, **kwargs) ^^^^^^^^^^^^^^^^^^^^^ File "/mmfs1/gscratch/dirac/shared/opt/conda/envs/lsst-scipipe-8.0.0/lib/python3.11/site-packages/parsl/addresses.py", line 93, in address_by_interface return socket.inet_ntoa(fcntl.ioctl( ^^^^^^^^^^^^ OSError: [Errno 99] Cannot assign requested address lsst.ctrl.bps.parsl INFO: Writing workflow with ID=/mmfs1/home/stevengs/dirac/DEEP/submit/DEEP/20190601/A1a/science#step1/20240313T015456Z/submit/DEEP/20190601/A1a/science#step1/20240313T015456Z lsst.ctrl.bps.drivers INFO: Prepare stage completed: Took 0.1001 seconds; current memory usage: 0.338 Gibyte, delta: 0.001 Gibyte, peak delta: 0.001 Gibyte lsst.ctrl.bps.drivers INFO: Peak memory usage for bps process 0.338 Gibyte (main), 0.325 Gibyte (largest child process) lsst.ctrl.bps.drivers INFO: Starting submit stage lsst.ctrl.bps.submit INFO: Submitting run to a workflow management system for execution parsl.dataflow.rundirs DEBUG: Parsl run initializing in rundir: runinfo/000 parsl.dataflow.dflow INFO: Starting DataFlowKernel with config Config( app_cache=True, checkpoint_files=None, checkpoint_mode='task_exit', checkpoint_period=None, executors=[MultiHighThroughputExecutor()], garbage_collect=True, initialize_logging=True, internal_tasks_max_threads=10, max_idletime=120.0, monitoring=None, retries=1, retry_handler=None, run_dir='runinfo', strategy='simple', usage_tracking=False ) parsl.dataflow.dflow INFO: Parsl version: 2023.06.12 parsl.usage_tracking.usage DEBUG: Tracking status: False parsl.dataflow.dflow INFO: Run id is: b7bd230f-d60b-455d-9d4c-a6e719240d4b parsl.dataflow.dflow DEBUG: Considering candidate for workflow name: /mmfs1/gscratch/dirac/shared/opt/conda/envs/lsst-scipipe-8.0.0/lib/python3.11/site-packages/parsl/dataflow/dflow.py parsl.dataflow.dflow DEBUG: Considering candidate for workflow name: /mmfs1/gscratch/dirac/shared/opt/conda/envs/lsst-scipipe-8.0.0/lib/python3.11/site-packages/typeguard/__init__.py parsl.dataflow.dflow DEBUG: Considering candidate for workflow name: /mmfs1/gscratch/dirac/shared/opt/conda/envs/lsst-scipipe-8.0.0/lib/python3.11/site-packages/parsl/dataflow/dflow.py parsl.dataflow.dflow DEBUG: Considering candidate for workflow name: /mmfs1/gscratch/dirac/shared/opt/conda/envs/lsst-scipipe-8.0.0/lib/python3.11/site-packages/typeguard/__init__.py parsl.dataflow.dflow DEBUG: Considering candidate for workflow name: /mmfs1/gscratch/dirac/shared/opt/conda/envs/lsst-scipipe-8.0.0/share/eups/Linux64/ctrl_bps_parsl/g145af14111+5b908e21bc/python/lsst/ctrl/bps/parsl/workflow.py parsl.dataflow.dflow DEBUG: Using workflow.py as workflow name parsl.dataflow.memoization INFO: App caching initialized parsl.dataflow.strategy DEBUG: Scaling strategy: simple parsl.executors.high_throughput.executor DEBUG: Starting queue management thread parsl.executors.high_throughput.executor DEBUG: queue management worker starting parsl.executors.high_throughput.executor DEBUG: Started queue management thread Submit dir: /mmfs1/home/stevengs/dirac/DEEP/submit/DEEP/20190601/A1a/science#step1/20240313T015456Z/submit/DEEP/20190601/A1a/science#step1/20240313T015456Z parsl.executors.high_throughput.executor DEBUG: Created management thread: parsl.executors.high_throughput.executor DEBUG: Launch command: process_worker_pool.py -a n3009,169.254.95.120,198.48.92.26,127.0.0.1,10.64.65.9,10.64.129.9 -p 0 -c 1.0 -m None --poll 10 --task_port=54190 --result_port=54623 --logdir=/mmfs1/gscratch/dirac/DEEP/submit/DEEP/20190601/A1a/science#step1/20240313T015456Z/runinfo/000/multi --block_id={block_id} --hb_period=30 --hb_threshold=120 --cpu-affinity none --available-accelerators --start-method spawn parsl.executors.high_throughput.executor DEBUG: Starting HighThroughputExecutor with provider: parsl.executors.status_handling INFO: Scaling out by 1 blocks parsl.executors.status_handling INFO: Allocated block ID 0 parsl.executors.status_handling DEBUG: Submitting to provider with job_name parsl.multi.block-0 2024-03-13 01:55:32 proc_lsst.multi:153 [INFO] [multi] got submit process_worker_pool.py -a n3009,169.254.95.120,198.48.92.26,127.0.0.1,10.64.65.9,10.64.129.9 -p 0 -c 1.0 -m None --poll 10 --task_port=54190 --result_port=54623 --logdir=/mmfs1/gscratch/dirac/DEEP/submit/DEEP/20190601/A1a/science#step1/20240313T015456Z/runinfo/000/multi --block_id=0 --hb_period=30 --hb_threshold=120 --cpu-affinity none --available-accelerators --start-method spawn 1 parsl.multi.block-0 proc_lsst.multi INFO: [multi] got submit process_worker_pool.py -a n3009,169.254.95.120,198.48.92.26,127.0.0.1,10.64.65.9,10.64.129.9 -p 0 -c 1.0 -m None --poll 10 --task_port=54190 --result_port=54623 --logdir=/mmfs1/gscratch/dirac/DEEP/submit/DEEP/20190601/A1a/science#step1/20240313T015456Z/runinfo/000/multi --block_id=0 --hb_period=30 --hb_threshold=120 --cpu-affinity none --available-accelerators --start-method spawn 1 parsl.multi.block-0 2024-03-13 01:55:32 proc_lsst.multi:162 [INFO] [multi] local process_worker_pool.py -a n3009,169.254.95.120,198.48.92.26,127.0.0.1,10.64.65.9,10.64.129.9 -p 0 -c 1.0 -m None --poll 10 --task_port=54190 --result_port=54623 --logdir=/mmfs1/gscratch/dirac/DEEP/submit/DEEP/20190601/A1a/science#step1/20240313T015456Z/runinfo/000/multi --block_id=0 --hb_period=30 --hb_threshold=120 --cpu-affinity none --available-accelerators --start-method spawn --max_workers 2 proc_lsst.multi INFO: [multi] local process_worker_pool.py -a n3009,169.254.95.120,198.48.92.26,127.0.0.1,10.64.65.9,10.64.129.9 -p 0 -c 1.0 -m None --poll 10 --task_port=54190 --result_port=54623 --logdir=/mmfs1/gscratch/dirac/DEEP/submit/DEEP/20190601/A1a/science#step1/20240313T015456Z/runinfo/000/multi --block_id=0 --hb_period=30 --hb_threshold=120 --cpu-affinity none --available-accelerators --start-method spawn --max_workers 2 2024-03-13 01:55:32 proc_lsst.multi:163 [INFO] [multi] len(self.providers[provider].resources) 0 proc_lsst.multi INFO: [multi] len(self.providers[provider].resources) 0 2024-03-13 01:55:32 proc_lsst.multi:164 [INFO] [multi] self.providers[provider].max_blocks 1 proc_lsst.multi INFO: [multi] self.providers[provider].max_blocks 1 2024-03-13 01:55:32 proc_lsst.multi:166 [INFO] [multi] submitting process_worker_pool.py -a n3009,169.254.95.120,198.48.92.26,127.0.0.1,10.64.65.9,10.64.129.9 -p 0 -c 1.0 -m None --poll 10 --task_port=54190 --result_port=54623 --logdir=/mmfs1/gscratch/dirac/DEEP/submit/DEEP/20190601/A1a/science#step1/20240313T015456Z/runinfo/000/multi --block_id=0 --hb_period=30 --hb_threshold=120 --cpu-affinity none --available-accelerators --start-method spawn --max_workers 2 to local proc_lsst.multi INFO: [multi] submitting process_worker_pool.py -a n3009,169.254.95.120,198.48.92.26,127.0.0.1,10.64.65.9,10.64.129.9 -p 0 -c 1.0 -m None --poll 10 --task_port=54190 --result_port=54623 --logdir=/mmfs1/gscratch/dirac/DEEP/submit/DEEP/20190601/A1a/science#step1/20240313T015456Z/runinfo/000/multi --block_id=0 --hb_period=30 --hb_threshold=120 --cpu-affinity none --available-accelerators --start-method spawn --max_workers 2 to local parsl.providers.local.local DEBUG: Launching in remote mode 2024-03-13 01:55:32 proc_lsst.multi:170 [INFO] [multi] job_id 30278 proc_lsst.multi INFO: [multi] job_id 30278 2024-03-13 01:55:32 proc_lsst.multi:171 [INFO] [multi] len(self.providers[provider].resources) 1 proc_lsst.multi INFO: [multi] len(self.providers[provider].resources) 1 2024-03-13 01:55:32 proc_lsst.multi:172 [INFO] [multi] self.providers[provider].max_blocks 1 proc_lsst.multi INFO: [multi] self.providers[provider].max_blocks 1 2024-03-13 01:55:32 proc_lsst.multi:178 [INFO] [multi] provider local accepted submit and returned 30278 proc_lsst.multi INFO: [multi] provider local accepted submit and returned 30278 parsl.executors.status_handling DEBUG: Launched block 0 on executor multi with job ID 30278 parsl.dataflow.job_status_poller DEBUG: Adding executor multi parsl.dataflow.dflow DEBUG: Task 0 will be sent to executor multi parsl.dataflow.dflow DEBUG: Adding output dependencies parsl.dataflow.dflow INFO: Task 0 submitted for App characterizeImage, not waiting on any dependency parsl.dataflow.dflow DEBUG: Task 0 set to pending state with AppFuture: parsl.dataflow.memoization DEBUG: Ignoring these kwargs for checkpointing: ['stderr', 'stdout'] parsl.dataflow.memoization DEBUG: Ignoring kwarg stderr parsl.dataflow.memoization DEBUG: Ignoring kwarg stdout parsl.dataflow.memoization DEBUG: Task 0 has memoization hash 3200452dca1a7d7480267654caca544e parsl.dataflow.memoization INFO: Task 0 had no result in cache parsl.executors.high_throughput.executor DEBUG: Pushing function .wrapper at 0x146087350900> to queue with args ("'${CTRL_MPEXEC_DIR}/bin/pipetask --long-log --log-level=VERBOSE run-qbb /mmfs1/home/stevengs/dirac/D...",) parsl.dataflow.dflow INFO: Parsl task 0 try 0 launched on executor multi with executor id 1 parsl.dataflow.dflow INFO: Standard output for task 0 available at /mmfs1/home/stevengs/dirac/DEEP/submit/DEEP/20190601/A1a/science#step1/20240313T015456Z/submit/DEEP/20190601/A1a/science#step1/20240313T015456Z/logs/characterizeImage/860799/9adb4454-aa50-46a1-be9c-1914d10d8192_characterizeImage_860799_2.stdout parsl.dataflow.dflow INFO: Standard error for task 0 available at /mmfs1/home/stevengs/dirac/DEEP/submit/DEEP/20190601/A1a/science#step1/20240313T015456Z/submit/DEEP/20190601/A1a/science#step1/20240313T015456Z/logs/characterizeImage/860799/9adb4454-aa50-46a1-be9c-1914d10d8192_characterizeImage_860799_2.stderr parsl.dataflow.dflow DEBUG: Task 1 will be sent to executor multi parsl.dataflow.dflow DEBUG: Adding output dependencies parsl.dataflow.dflow INFO: Task 1 submitted for App calibrate, waiting on task 0 parsl.dataflow.dflow DEBUG: Task 1 set to pending state with AppFuture: parsl.dataflow.dflow DEBUG: Task 1 has outstanding dependencies, so launch_if_ready skipping parsl.dataflow.dflow DEBUG: Task 2 will be sent to executor multi parsl.dataflow.dflow DEBUG: Adding output dependencies parsl.dataflow.dflow INFO: Task 2 submitted for App characterizeImage, not waiting on any dependency parsl.dataflow.dflow DEBUG: Task 2 set to pending state with AppFuture: parsl.dataflow.memoization DEBUG: Ignoring these kwargs for checkpointing: ['stderr', 'stdout'] parsl.dataflow.memoization DEBUG: Ignoring kwarg stderr parsl.dataflow.memoization DEBUG: Ignoring kwarg stdout parsl.dataflow.memoization DEBUG: Task 2 has memoization hash bb5e55c0a2be46ba21469a3993c8751c parsl.dataflow.memoization INFO: Task 2 had no result in cache parsl.executors.high_throughput.executor DEBUG: Pushing function .wrapper at 0x146087351080> to queue with args ("'${CTRL_MPEXEC_DIR}/bin/pipetask --long-log --log-level=VERBOSE run-qbb /mmfs1/home/stevengs/dirac/D...",) parsl.dataflow.dflow INFO: Parsl task 2 try 0 launched on executor multi with executor id 2 parsl.dataflow.dflow INFO: Standard output for task 2 available at /mmfs1/home/stevengs/dirac/DEEP/submit/DEEP/20190601/A1a/science#step1/20240313T015456Z/submit/DEEP/20190601/A1a/science#step1/20240313T015456Z/logs/characterizeImage/860800/2d624bbd-9726-4837-bcb8-41adea972484_characterizeImage_860800_2.stdout parsl.dataflow.dflow INFO: Standard error for task 2 available at /mmfs1/home/stevengs/dirac/DEEP/submit/DEEP/20190601/A1a/science#step1/20240313T015456Z/submit/DEEP/20190601/A1a/science#step1/20240313T015456Z/logs/characterizeImage/860800/2d624bbd-9726-4837-bcb8-41adea972484_characterizeImage_860800_2.stderr parsl.dataflow.dflow DEBUG: Task 3 will be sent to executor multi parsl.dataflow.dflow DEBUG: Adding output dependencies parsl.dataflow.dflow INFO: Task 3 submitted for App calibrate, waiting on task 2 parsl.dataflow.dflow DEBUG: Task 3 set to pending state with AppFuture: parsl.dataflow.dflow DEBUG: Task 3 has outstanding dependencies, so launch_if_ready skipping parsl.dataflow.dflow DEBUG: Task 4 will be sent to executor multi parsl.dataflow.dflow DEBUG: Adding output dependencies parsl.dataflow.dflow INFO: Task 4 submitted for App characterizeImage, not waiting on any dependency parsl.dataflow.dflow DEBUG: Task 4 set to pending state with AppFuture: parsl.dataflow.memoization DEBUG: Ignoring these kwargs for checkpointing: ['stderr', 'stdout'] parsl.dataflow.memoization DEBUG: Ignoring kwarg stderr parsl.dataflow.memoization DEBUG: Ignoring kwarg stdout parsl.dataflow.memoization DEBUG: Task 4 has memoization hash a6354c1823bd97af85b36decee46236d parsl.dataflow.memoization INFO: Task 4 had no result in cache parsl.executors.high_throughput.executor DEBUG: Pushing function .wrapper at 0x146087350fe0> to queue with args ("'${CTRL_MPEXEC_DIR}/bin/pipetask --long-log --log-level=VERBOSE run-qbb /mmfs1/home/stevengs/dirac/D...",) parsl.dataflow.dflow INFO: Parsl task 4 try 0 launched on executor multi with executor id 3 parsl.dataflow.dflow INFO: Standard output for task 4 available at /mmfs1/home/stevengs/dirac/DEEP/submit/DEEP/20190601/A1a/science#step1/20240313T015456Z/submit/DEEP/20190601/A1a/science#step1/20240313T015456Z/logs/characterizeImage/860798/b922026b-ba89-413a-b870-8157195c258c_characterizeImage_860798_2.stdout parsl.dataflow.dflow INFO: Standard error for task 4 available at /mmfs1/home/stevengs/dirac/DEEP/submit/DEEP/20190601/A1a/science#step1/20240313T015456Z/submit/DEEP/20190601/A1a/science#step1/20240313T015456Z/logs/characterizeImage/860798/b922026b-ba89-413a-b870-8157195c258c_characterizeImage_860798_2.stderr parsl.dataflow.dflow DEBUG: Task 5 will be sent to executor multi parsl.dataflow.dflow DEBUG: Adding output dependencies parsl.dataflow.dflow INFO: Task 5 submitted for App calibrate, waiting on task 4 parsl.dataflow.dflow DEBUG: Task 5 set to pending state with AppFuture: parsl.dataflow.dflow DEBUG: Task 5 has outstanding dependencies, so launch_if_ready skipping 2024-03-13 01:55:33 proc_lsst.multi:146 [INFO] found job 30278 in provider local proc_lsst.multi INFO: found job 30278 in provider local parsl.dataflow.strategy DEBUG: general strategy starting with strategy_type simple for 1 executors parsl.dataflow.strategy DEBUG: Strategizing for executor multi parsl.dataflow.strategy DEBUG: Slot ratio calculation: active_slots = 1, active_tasks = 3 parsl.dataflow.strategy DEBUG: Executor multi has 3 active tasks, 1/0 running/pending blocks, and 0 connected workers parsl.dataflow.strategy DEBUG: Strategy case 3: no changes necessary to current block load parsl.process_loggers DEBUG: Normal ending for _general_strategy on thread JobStatusPoller-Timer-Thread-22404838064272 parsl.dataflow.strategy DEBUG: general strategy starting with strategy_type simple for 1 executors parsl.dataflow.strategy DEBUG: Strategizing for executor multi parsl.dataflow.strategy DEBUG: Slot ratio calculation: active_slots = 1, active_tasks = 3 parsl.dataflow.strategy DEBUG: Executor multi has 3 active tasks, 1/0 running/pending blocks, and 2 connected workers parsl.dataflow.strategy DEBUG: Strategy case 3: no changes necessary to current block load parsl.process_loggers DEBUG: Normal ending for _general_strategy on thread JobStatusPoller-Timer-Thread-22404838064272 parsl.app.errors DEBUG: Reraising exception of type parsl.dataflow.dflow DEBUG: Task 2 try 0 failed parsl.dataflow.dflow INFO: Task 2 marked for retry parsl.dataflow.dflow INFO: Standard output for task 2 available at /mmfs1/home/stevengs/dirac/DEEP/submit/DEEP/20190601/A1a/science#step1/20240313T015456Z/submit/DEEP/20190601/A1a/science#step1/20240313T015456Z/logs/characterizeImage/860800/2d624bbd-9726-4837-bcb8-41adea972484_characterizeImage_860800_2.stdout parsl.dataflow.dflow INFO: Standard error for task 2 available at /mmfs1/home/stevengs/dirac/DEEP/submit/DEEP/20190601/A1a/science#step1/20240313T015456Z/submit/DEEP/20190601/A1a/science#step1/20240313T015456Z/logs/characterizeImage/860800/2d624bbd-9726-4837-bcb8-41adea972484_characterizeImage_860800_2.stderr parsl.dataflow.memoization DEBUG: Ignoring these kwargs for checkpointing: ['stderr', 'stdout'] parsl.dataflow.memoization DEBUG: Ignoring kwarg stderr parsl.dataflow.memoization DEBUG: Ignoring kwarg stdout parsl.dataflow.memoization DEBUG: Task 2 has memoization hash bb5e55c0a2be46ba21469a3993c8751c parsl.dataflow.memoization INFO: Task 2 had no result in cache parsl.executors.high_throughput.executor DEBUG: Pushing function .wrapper at 0x146087351080> to queue with args ("'${CTRL_MPEXEC_DIR}/bin/pipetask --long-log --log-level=VERBOSE run-qbb /mmfs1/home/stevengs/dirac/D...",) parsl.dataflow.dflow INFO: Parsl task 2 try 1 launched on executor multi with executor id 4 parsl.dataflow.dflow INFO: Standard output for task 2 available at /mmfs1/home/stevengs/dirac/DEEP/submit/DEEP/20190601/A1a/science#step1/20240313T015456Z/submit/DEEP/20190601/A1a/science#step1/20240313T015456Z/logs/characterizeImage/860800/2d624bbd-9726-4837-bcb8-41adea972484_characterizeImage_860800_2.stdout parsl.dataflow.dflow INFO: Standard error for task 2 available at /mmfs1/home/stevengs/dirac/DEEP/submit/DEEP/20190601/A1a/science#step1/20240313T015456Z/submit/DEEP/20190601/A1a/science#step1/20240313T015456Z/logs/characterizeImage/860800/2d624bbd-9726-4837-bcb8-41adea972484_characterizeImage_860800_2.stderr parsl.app.errors DEBUG: Reraising exception of type parsl.dataflow.dflow DEBUG: Task 0 try 0 failed parsl.dataflow.dflow INFO: Task 0 marked for retry parsl.dataflow.dflow INFO: Standard output for task 0 available at /mmfs1/home/stevengs/dirac/DEEP/submit/DEEP/20190601/A1a/science#step1/20240313T015456Z/submit/DEEP/20190601/A1a/science#step1/20240313T015456Z/logs/characterizeImage/860799/9adb4454-aa50-46a1-be9c-1914d10d8192_characterizeImage_860799_2.stdout parsl.dataflow.dflow INFO: Standard error for task 0 available at /mmfs1/home/stevengs/dirac/DEEP/submit/DEEP/20190601/A1a/science#step1/20240313T015456Z/submit/DEEP/20190601/A1a/science#step1/20240313T015456Z/logs/characterizeImage/860799/9adb4454-aa50-46a1-be9c-1914d10d8192_characterizeImage_860799_2.stderr parsl.dataflow.memoization DEBUG: Ignoring these kwargs for checkpointing: ['stderr', 'stdout'] parsl.dataflow.memoization DEBUG: Ignoring kwarg stderr parsl.dataflow.memoization DEBUG: Ignoring kwarg stdout parsl.dataflow.memoization DEBUG: Task 0 has memoization hash 3200452dca1a7d7480267654caca544e parsl.dataflow.memoization INFO: Task 0 had no result in cache parsl.executors.high_throughput.executor DEBUG: Pushing function .wrapper at 0x146087350900> to queue with args ("'${CTRL_MPEXEC_DIR}/bin/pipetask --long-log --log-level=VERBOSE run-qbb /mmfs1/home/stevengs/dirac/D...",) parsl.dataflow.dflow INFO: Parsl task 0 try 1 launched on executor multi with executor id 5 parsl.dataflow.dflow INFO: Standard output for task 0 available at /mmfs1/home/stevengs/dirac/DEEP/submit/DEEP/20190601/A1a/science#step1/20240313T015456Z/submit/DEEP/20190601/A1a/science#step1/20240313T015456Z/logs/characterizeImage/860799/9adb4454-aa50-46a1-be9c-1914d10d8192_characterizeImage_860799_2.stdout parsl.dataflow.dflow INFO: Standard error for task 0 available at /mmfs1/home/stevengs/dirac/DEEP/submit/DEEP/20190601/A1a/science#step1/20240313T015456Z/submit/DEEP/20190601/A1a/science#step1/20240313T015456Z/logs/characterizeImage/860799/9adb4454-aa50-46a1-be9c-1914d10d8192_characterizeImage_860799_2.stderr parsl.dataflow.strategy DEBUG: general strategy starting with strategy_type simple for 1 executors parsl.dataflow.strategy DEBUG: Strategizing for executor multi parsl.dataflow.strategy DEBUG: Slot ratio calculation: active_slots = 1, active_tasks = 3 parsl.dataflow.strategy DEBUG: Executor multi has 3 active tasks, 1/0 running/pending blocks, and 2 connected workers parsl.dataflow.strategy DEBUG: Strategy case 3: no changes necessary to current block load parsl.process_loggers DEBUG: Normal ending for _general_strategy on thread JobStatusPoller-Timer-Thread-22404838064272 parsl.dataflow.strategy DEBUG: general strategy starting with strategy_type simple for 1 executors parsl.dataflow.strategy DEBUG: Strategizing for executor multi parsl.dataflow.strategy DEBUG: Slot ratio calculation: active_slots = 1, active_tasks = 3 parsl.dataflow.strategy DEBUG: Executor multi has 3 active tasks, 1/0 running/pending blocks, and 2 connected workers parsl.dataflow.strategy DEBUG: Strategy case 3: no changes necessary to current block load parsl.process_loggers DEBUG: Normal ending for _general_strategy on thread JobStatusPoller-Timer-Thread-22404838064272 parsl.app.errors DEBUG: Reraising exception of type parsl.dataflow.dflow DEBUG: Task 4 try 0 failed parsl.dataflow.dflow INFO: Task 4 marked for retry parsl.dataflow.dflow INFO: Standard output for task 4 available at /mmfs1/home/stevengs/dirac/DEEP/submit/DEEP/20190601/A1a/science#step1/20240313T015456Z/submit/DEEP/20190601/A1a/science#step1/20240313T015456Z/logs/characterizeImage/860798/b922026b-ba89-413a-b870-8157195c258c_characterizeImage_860798_2.stdout parsl.dataflow.dflow INFO: Standard error for task 4 available at /mmfs1/home/stevengs/dirac/DEEP/submit/DEEP/20190601/A1a/science#step1/20240313T015456Z/submit/DEEP/20190601/A1a/science#step1/20240313T015456Z/logs/characterizeImage/860798/b922026b-ba89-413a-b870-8157195c258c_characterizeImage_860798_2.stderr parsl.dataflow.memoization DEBUG: Ignoring these kwargs for checkpointing: ['stderr', 'stdout'] parsl.dataflow.memoization DEBUG: Ignoring kwarg stderr parsl.dataflow.memoization DEBUG: Ignoring kwarg stdout parsl.dataflow.memoization DEBUG: Task 4 has memoization hash a6354c1823bd97af85b36decee46236d parsl.dataflow.memoization INFO: Task 4 had no result in cache parsl.executors.high_throughput.executor DEBUG: Pushing function .wrapper at 0x146087350fe0> to queue with args ("'${CTRL_MPEXEC_DIR}/bin/pipetask --long-log --log-level=VERBOSE run-qbb /mmfs1/home/stevengs/dirac/D...",) parsl.dataflow.dflow INFO: Parsl task 4 try 1 launched on executor multi with executor id 6 parsl.dataflow.dflow INFO: Standard output for task 4 available at /mmfs1/home/stevengs/dirac/DEEP/submit/DEEP/20190601/A1a/science#step1/20240313T015456Z/submit/DEEP/20190601/A1a/science#step1/20240313T015456Z/logs/characterizeImage/860798/b922026b-ba89-413a-b870-8157195c258c_characterizeImage_860798_2.stdout parsl.dataflow.dflow INFO: Standard error for task 4 available at /mmfs1/home/stevengs/dirac/DEEP/submit/DEEP/20190601/A1a/science#step1/20240313T015456Z/submit/DEEP/20190601/A1a/science#step1/20240313T015456Z/logs/characterizeImage/860798/b922026b-ba89-413a-b870-8157195c258c_characterizeImage_860798_2.stderr parsl.app.errors DEBUG: Reraising exception of type parsl.dataflow.dflow DEBUG: Task 2 try 1 failed parsl.dataflow.dflow ERROR: Task 2 failed after 1 retry attempts Traceback (most recent call last): File "/mmfs1/gscratch/dirac/shared/opt/conda/envs/lsst-scipipe-8.0.0/lib/python3.11/site-packages/parsl/dataflow/dflow.py", line 300, in handle_exec_update res = self._unwrap_remote_exception_wrapper(future) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/mmfs1/gscratch/dirac/shared/opt/conda/envs/lsst-scipipe-8.0.0/lib/python3.11/site-packages/parsl/dataflow/dflow.py", line 566, in _unwrap_remote_exception_wrapper result.reraise() File "/mmfs1/gscratch/dirac/shared/opt/conda/envs/lsst-scipipe-8.0.0/lib/python3.11/site-packages/parsl/app/errors.py", line 123, in reraise reraise(t, v, v.__traceback__) File "/mmfs1/gscratch/dirac/shared/opt/conda/envs/lsst-scipipe-8.0.0/lib/python3.11/site-packages/six.py", line 719, in reraise raise value File "/mmfs1/gscratch/dirac/shared/opt/conda/envs/lsst-scipipe-8.0.0/lib/python3.11/site-packages/parsl/app/errors.py", line 146, in wrapper return func(*args, **kwargs) ^^^^^^^^^^^^^^^^^ File "/mmfs1/gscratch/dirac/shared/opt/conda/envs/lsst-scipipe-8.0.0/lib/python3.11/site-packages/parsl/app/bash.py", line 86, in remote_side_bash_executor raise pe.BashExitFailure(func_name, proc.returncode) ^^^^^^^^^^^^^^^^^ parsl.app.errors.BashExitFailure: bash_app characterizeImage failed with unix exit code 1 parsl.dataflow.memoization DEBUG: Storing app cache entry bb5e55c0a2be46ba21469a3993c8751c with result from task 2 parsl.dataflow.dflow WARNING: No tasks checkpointed so far in this run. Please ensure caching is enabled parsl.dataflow.dflow INFO: Task 3 failed due to dependency failure parsl.dataflow.dflow DEBUG: Task 3 try 0 failed parsl.dataflow.dflow INFO: Task 3 failed due to dependency failure so skipping retries parsl.dataflow.memoization ERROR: Attempting to update app cache entry but hashsum is not a string key parsl.dataflow.dflow WARNING: No tasks checkpointed so far in this run. Please ensure caching is enabled parsl.dataflow.dflow INFO: Standard output for task 3 available at /mmfs1/home/stevengs/dirac/DEEP/submit/DEEP/20190601/A1a/science#step1/20240313T015456Z/submit/DEEP/20190601/A1a/science#step1/20240313T015456Z/logs/calibrate/860800/b28ad0cc-2d0a-468f-b302-2515c65f8b3e_calibrate_860800_2.stdout parsl.dataflow.dflow INFO: Standard error for task 3 available at /mmfs1/home/stevengs/dirac/DEEP/submit/DEEP/20190601/A1a/science#step1/20240313T015456Z/submit/DEEP/20190601/A1a/science#step1/20240313T015456Z/logs/calibrate/860800/b28ad0cc-2d0a-468f-b302-2515c65f8b3e_calibrate_860800_2.stderr parsl.dataflow.dflow INFO: Standard output for task 2 available at /mmfs1/home/stevengs/dirac/DEEP/submit/DEEP/20190601/A1a/science#step1/20240313T015456Z/submit/DEEP/20190601/A1a/science#step1/20240313T015456Z/logs/characterizeImage/860800/2d624bbd-9726-4837-bcb8-41adea972484_characterizeImage_860800_2.stdout parsl.dataflow.dflow INFO: Standard error for task 2 available at /mmfs1/home/stevengs/dirac/DEEP/submit/DEEP/20190601/A1a/science#step1/20240313T015456Z/submit/DEEP/20190601/A1a/science#step1/20240313T015456Z/logs/characterizeImage/860800/2d624bbd-9726-4837-bcb8-41adea972484_characterizeImage_860800_2.stderr parsl.dataflow.strategy DEBUG: general strategy starting with strategy_type simple for 1 executors parsl.dataflow.strategy DEBUG: Strategizing for executor multi parsl.dataflow.strategy DEBUG: Slot ratio calculation: active_slots = 1, active_tasks = 2 parsl.dataflow.strategy DEBUG: Executor multi has 2 active tasks, 1/0 running/pending blocks, and 2 connected workers parsl.dataflow.strategy DEBUG: Strategy case 3: no changes necessary to current block load parsl.process_loggers DEBUG: Normal ending for _general_strategy on thread JobStatusPoller-Timer-Thread-22404838064272 parsl.app.errors DEBUG: Reraising exception of type parsl.dataflow.dflow DEBUG: Task 0 try 1 failed parsl.dataflow.dflow ERROR: Task 0 failed after 1 retry attempts Traceback (most recent call last): File "/mmfs1/gscratch/dirac/shared/opt/conda/envs/lsst-scipipe-8.0.0/lib/python3.11/site-packages/parsl/dataflow/dflow.py", line 300, in handle_exec_update res = self._unwrap_remote_exception_wrapper(future) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/mmfs1/gscratch/dirac/shared/opt/conda/envs/lsst-scipipe-8.0.0/lib/python3.11/site-packages/parsl/dataflow/dflow.py", line 566, in _unwrap_remote_exception_wrapper result.reraise() File "/mmfs1/gscratch/dirac/shared/opt/conda/envs/lsst-scipipe-8.0.0/lib/python3.11/site-packages/parsl/app/errors.py", line 123, in reraise reraise(t, v, v.__traceback__) File "/mmfs1/gscratch/dirac/shared/opt/conda/envs/lsst-scipipe-8.0.0/lib/python3.11/site-packages/six.py", line 719, in reraise raise value File "/mmfs1/gscratch/dirac/shared/opt/conda/envs/lsst-scipipe-8.0.0/lib/python3.11/site-packages/parsl/app/errors.py", line 146, in wrapper return func(*args, **kwargs) ^^^^^^^^^^^^^^^^^ File "/mmfs1/gscratch/dirac/shared/opt/conda/envs/lsst-scipipe-8.0.0/lib/python3.11/site-packages/parsl/app/bash.py", line 86, in remote_side_bash_executor raise pe.BashExitFailure(func_name, proc.returncode) ^^^^^^^^^^^^^^^^^ parsl.app.errors.BashExitFailure: bash_app characterizeImage failed with unix exit code 1 parsl.dataflow.memoization DEBUG: Storing app cache entry 3200452dca1a7d7480267654caca544e with result from task 0 parsl.dataflow.dflow WARNING: No tasks checkpointed so far in this run. Please ensure caching is enabled parsl.dataflow.dflow INFO: Task 1 failed due to dependency failure parsl.dataflow.dflow DEBUG: Task 1 try 0 failed parsl.dataflow.dflow INFO: Task 1 failed due to dependency failure so skipping retries parsl.dataflow.memoization ERROR: Attempting to update app cache entry but hashsum is not a string key parsl.dataflow.dflow WARNING: No tasks checkpointed so far in this run. Please ensure caching is enabled parsl.dataflow.dflow INFO: Standard output for task 1 available at /mmfs1/home/stevengs/dirac/DEEP/submit/DEEP/20190601/A1a/science#step1/20240313T015456Z/submit/DEEP/20190601/A1a/science#step1/20240313T015456Z/logs/calibrate/860799/cf26fc6b-f2e5-41b0-9022-d3dea422c242_calibrate_860799_2.stdout parsl.dataflow.dflow INFO: Standard error for task 1 available at /mmfs1/home/stevengs/dirac/DEEP/submit/DEEP/20190601/A1a/science#step1/20240313T015456Z/submit/DEEP/20190601/A1a/science#step1/20240313T015456Z/logs/calibrate/860799/cf26fc6b-f2e5-41b0-9022-d3dea422c242_calibrate_860799_2.stderr parsl.dataflow.dflow INFO: Standard output for task 0 available at /mmfs1/home/stevengs/dirac/DEEP/submit/DEEP/20190601/A1a/science#step1/20240313T015456Z/submit/DEEP/20190601/A1a/science#step1/20240313T015456Z/logs/characterizeImage/860799/9adb4454-aa50-46a1-be9c-1914d10d8192_characterizeImage_860799_2.stdout parsl.dataflow.dflow INFO: Standard error for task 0 available at /mmfs1/home/stevengs/dirac/DEEP/submit/DEEP/20190601/A1a/science#step1/20240313T015456Z/submit/DEEP/20190601/A1a/science#step1/20240313T015456Z/logs/characterizeImage/860799/9adb4454-aa50-46a1-be9c-1914d10d8192_characterizeImage_860799_2.stderr parsl.app.errors DEBUG: Reraising exception of type parsl.dataflow.dflow DEBUG: Task 4 try 1 failed parsl.dataflow.dflow ERROR: Task 4 failed after 1 retry attempts Traceback (most recent call last): File "/mmfs1/gscratch/dirac/shared/opt/conda/envs/lsst-scipipe-8.0.0/lib/python3.11/site-packages/parsl/dataflow/dflow.py", line 300, in handle_exec_update res = self._unwrap_remote_exception_wrapper(future) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/mmfs1/gscratch/dirac/shared/opt/conda/envs/lsst-scipipe-8.0.0/lib/python3.11/site-packages/parsl/dataflow/dflow.py", line 566, in _unwrap_remote_exception_wrapper result.reraise() File "/mmfs1/gscratch/dirac/shared/opt/conda/envs/lsst-scipipe-8.0.0/lib/python3.11/site-packages/parsl/app/errors.py", line 123, in reraise reraise(t, v, v.__traceback__) File "/mmfs1/gscratch/dirac/shared/opt/conda/envs/lsst-scipipe-8.0.0/lib/python3.11/site-packages/six.py", line 719, in reraise raise value File "/mmfs1/gscratch/dirac/shared/opt/conda/envs/lsst-scipipe-8.0.0/lib/python3.11/site-packages/parsl/app/errors.py", line 146, in wrapper return func(*args, **kwargs) ^^^^^^^^^^^^^^^^^ File "/mmfs1/gscratch/dirac/shared/opt/conda/envs/lsst-scipipe-8.0.0/lib/python3.11/site-packages/parsl/app/bash.py", line 86, in remote_side_bash_executor raise pe.BashExitFailure(func_name, proc.returncode) ^^^^^^^^^^^^^^^^^ parsl.app.errors.BashExitFailure: bash_app characterizeImage failed with unix exit code 1 parsl.dataflow.memoization DEBUG: Storing app cache entry a6354c1823bd97af85b36decee46236d with result from task 4 parsl.dataflow.dflow WARNING: No tasks checkpointed so far in this run. Please ensure caching is enabled parsl.dataflow.dflow INFO: Task 5 failed due to dependency failure parsl.dataflow.dflow DEBUG: Task 5 try 0 failed parsl.dataflow.dflow INFO: Task 5 failed due to dependency failure so skipping retries parsl.dataflow.memoization ERROR: Attempting to update app cache entry but hashsum is not a string key parsl.dataflow.dflow INFO: DFK cleanup initiated parsl.dataflow.dflow INFO: Summary of tasks in DFK: parsl.dataflow.dflow INFO: Tasks in state States.unsched: 0 parsl.dataflow.dflow INFO: Tasks in state States.pending: 0 parsl.dataflow.dflow WARNING: No tasks checkpointed so far in this run. Please ensure caching is enabled parsl.dataflow.dflow INFO: Tasks in state States.running: 0 parsl.dataflow.dflow INFO: Standard output for task 5 available at /mmfs1/home/stevengs/dirac/DEEP/submit/DEEP/20190601/A1a/science#step1/20240313T015456Z/submit/DEEP/20190601/A1a/science#step1/20240313T015456Z/logs/calibrate/860798/2f554bdd-fd14-4790-80c9-3881c7fcf373_calibrate_860798_2.stdout parsl.dataflow.dflow INFO: Tasks in state States.exec_done: 0 parsl.dataflow.dflow INFO: Standard error for task 5 available at /mmfs1/home/stevengs/dirac/DEEP/submit/DEEP/20190601/A1a/science#step1/20240313T015456Z/submit/DEEP/20190601/A1a/science#step1/20240313T015456Z/logs/calibrate/860798/2f554bdd-fd14-4790-80c9-3881c7fcf373_calibrate_860798_2.stderr parsl.dataflow.dflow INFO: Tasks in state States.failed: 3 parsl.dataflow.dflow INFO: Standard output for task 4 available at /mmfs1/home/stevengs/dirac/DEEP/submit/DEEP/20190601/A1a/science#step1/20240313T015456Z/submit/DEEP/20190601/A1a/science#step1/20240313T015456Z/logs/characterizeImage/860798/b922026b-ba89-413a-b870-8157195c258c_characterizeImage_860798_2.stdout parsl.dataflow.dflow INFO: Tasks in state States.dep_fail: 3 parsl.dataflow.dflow INFO: Standard error for task 4 available at /mmfs1/home/stevengs/dirac/DEEP/submit/DEEP/20190601/A1a/science#step1/20240313T015456Z/submit/DEEP/20190601/A1a/science#step1/20240313T015456Z/logs/characterizeImage/860798/b922026b-ba89-413a-b870-8157195c258c_characterizeImage_860798_2.stderr parsl.dataflow.dflow INFO: Tasks in state States.launched: 0 parsl.dataflow.dflow INFO: Tasks in state States.fail_retryable: 0 parsl.dataflow.dflow INFO: Tasks in state States.memo_done: 0 parsl.dataflow.dflow INFO: Tasks in state States.joining: 0 parsl.dataflow.dflow INFO: Tasks in state States.running_ended: 0 parsl.dataflow.dflow INFO: End of summary parsl.dataflow.dflow WARNING: No tasks checkpointed so far in this run. Please ensure caching is enabled parsl.dataflow.dflow INFO: Closing job status poller parsl.dataflow.dflow INFO: Terminated job status poller parsl.dataflow.dflow INFO: Scaling in and shutting down executors parsl.dataflow.dflow INFO: Scaling in executor multi parsl.executors.high_throughput.executor DEBUG: Scale in called, blocks=1, block_ids=[] parsl.executors.high_throughput.executor DEBUG: Scale in selecting from 1 blocks parsl.executors.high_throughput.executor DEBUG: Sending hold to manager: 9e921e567b0d parsl.executors.high_throughput.executor DEBUG: Sent hold request to manager: 9e921e567b0d 2024-03-13 01:55:55 proc_lsst.multi:146 [INFO] found job 30278 in provider local proc_lsst.multi INFO: found job 30278 in provider local 2024-03-13 01:55:55 proc_lsst.multi:201 [INFO] cancelling 30278 on provider local proc_lsst.multi INFO: cancelling 30278 on provider local parsl.providers.local.local DEBUG: Terminating job/proc_id: 30278 parsl.dataflow.dflow INFO: Shutting down executor multi 2024-03-13 01:55:56 proc_lsst.multi:40 [INFO] Cancelling all provider resources proc_lsst.multi INFO: Cancelling all provider resources 2024-03-13 01:55:56 proc_lsst.multi:47 [INFO] new jobs since last cancel ['30278'] proc_lsst.multi INFO: new jobs since last cancel ['30278'] 2024-03-13 01:55:56 proc_lsst.multi:146 [INFO] found job 30278 in provider local proc_lsst.multi INFO: found job 30278 in provider local 2024-03-13 01:55:56 proc_lsst.multi:201 [INFO] cancelling 30278 on provider local proc_lsst.multi INFO: cancelling 30278 on provider local parsl.providers.local.local DEBUG: Terminating job/proc_id: 30278 parsl.providers.local.local WARNING: Failed to kill PID: 30278 and child processes on local 2024-03-13 01:55:57 proc_lsst.multi:50 [INFO] no new jobs since last cancel, resuming executor shutdown proc_lsst.multi INFO: no new jobs since last cancel, resuming executor shutdown parsl.executors.high_throughput.executor INFO: Attempting HighThroughputExecutor shutdown parsl.executors.high_throughput.executor INFO: Finished HighThroughputExecutor shutdown attempt parsl.dataflow.dflow INFO: Shut down executor multi parsl.dataflow.dflow INFO: Shutting down executor _parsl_internal parsl.executors.threads DEBUG: Shutting down executor, which involves waiting for running tasks to complete parsl.executors.threads DEBUG: Done with executor shutdown parsl.dataflow.dflow INFO: Shut down executor _parsl_internal parsl.dataflow.dflow INFO: Terminated executors parsl.dataflow.dflow INFO: DFK cleanup complete parsl.process_loggers DEBUG: Normal ending for cleanup on thread MainThread lsst.ctrl.bps.submit INFO: Completed submitting to a workflow management system: Took 42.4394 seconds lsst.ctrl.bps.drivers INFO: Run 'DEEP_20190601_A1a_science#step1_20240313T015456Z' submitted for execution with id 'None' lsst.ctrl.bps.drivers INFO: Completed submit stage: Took 42.4488 seconds; current memory usage: 0.339 Gibyte, delta: 0.001 Gibyte, peak delta: 0.001 Gibyte lsst.ctrl.bps.drivers INFO: Completed entire submission process: Took 57.7121 seconds; current memory usage: 0.339 Gibyte, delta: 0.202 Gibyte, peak delta: 0.202 Gibyte lsst.ctrl.bps.drivers INFO: Peak memory usage for bps process 0.339 Gibyte (main), 0.339 Gibyte (largest child process) Run Id: None Run Name: DEEP_20190601_A1a_science#step1_20240313T015456Z parsl.dataflow.dflow INFO: python process is exiting, but DFK has already been cleaned up real 1m0.590s user 0m30.184s sys 0m5.174s