3 """A simple utility to unpickle some code, run it, and pickle the
13 from typing import Optional
15 import cloudpickle # type: ignore
16 import psutil # type: ignore
21 from stopwatch import Timer
22 from thread_utils import background_thread
24 logger = logging.getLogger(__file__)
26 cfg = config.add_commandline_args(
27 f"Remote Worker ({__file__})",
28 "Helper to run pickled code remotely and return results",
35 help='The location of the bundle of code to execute.',
42 help='The location where we should write the computation results.',
46 action=argparse_utils.ActionNoYes,
48 help='Should we watch for the cancellation of our parent ssh process?',
53 def watch_for_cancel(terminate_event: threading.Event) -> None:
54 logger.debug('Starting up background thread...')
55 p = psutil.Process(os.getpid())
58 ancestors = p.parents()
59 for ancestor in ancestors:
60 name = ancestor.name()
62 logger.debug(f'Ancestor process {name} (pid={pid})')
63 if 'ssh' in name.lower():
68 'Did not see sshd in our ancestors list?! Committing suicide.'
71 os.kill(os.getpid(), signal.SIGTERM)
73 os.kill(os.getpid(), signal.SIGKILL)
75 if terminate_event.is_set():
81 thread: Optional[threading.Thread],
82 stop_thread: Optional[threading.Event],
85 if stop_thread is not None:
87 assert thread is not None
94 in_file = config.config['code_file']
95 out_file = config.config['result_file']
99 if config.config['watch_for_cancel']:
100 (thread, stop_thread) = watch_for_cancel()
102 logger.debug(f'Reading {in_file}.')
104 with open(in_file, 'rb') as rb:
105 serialized = rb.read()
106 except Exception as e:
108 logger.critical(f'Problem reading {in_file}. Aborting.')
109 cleanup_and_exit(thread, stop_thread, 1)
111 logger.debug(f'Deserializing {in_file}.')
113 fun, args, kwargs = cloudpickle.loads(serialized)
114 except Exception as e:
116 logger.critical(f'Problem deserializing {in_file}. Aborting.')
117 cleanup_and_exit(thread, stop_thread, 2)
119 logger.debug('Invoking user code...')
121 ret = fun(*args, **kwargs)
122 logger.debug(f'User code took {t():.1f}s')
124 logger.debug('Serializing results')
126 serialized = cloudpickle.dumps(ret)
127 except Exception as e:
129 logger.critical(f'Could not serialize result ({type(ret)}). Aborting.')
130 cleanup_and_exit(thread, stop_thread, 3)
132 logger.debug(f'Writing {out_file}.')
134 with open(out_file, 'wb') as wb:
136 except Exception as e:
138 logger.critical(f'Error writing {out_file}. Aborting.')
139 cleanup_and_exit(thread, stop_thread, 4)
140 cleanup_and_exit(thread, stop_thread, 0)
143 if __name__ == '__main__':