ominfra 0.0.0.dev76__py3-none-any.whl → 0.0.0.dev78__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -0,0 +1,3323 @@
1
+ #!/usr/bin/env python3
2
+ # noinspection DuplicatedCode
3
+ # @omlish-lite
4
+ # @omlish-script
5
+ # @omlish-amalg-output ../supervisor/supervisor.py
6
+ # ruff: noqa: N802 UP006 UP007 UP036
7
+ import abc
8
+ import contextlib
9
+ import dataclasses as dc
10
+ import datetime
11
+ import errno
12
+ import fcntl
13
+ import functools
14
+ import grp
15
+ import json
16
+ import logging
17
+ import os
18
+ import pwd
19
+ import re
20
+ import resource
21
+ import select
22
+ import shlex
23
+ import signal
24
+ import stat
25
+ import sys
26
+ import tempfile
27
+ import threading
28
+ import time
29
+ import traceback
30
+ import types
31
+ import typing as ta
32
+ import warnings
33
+
34
+
35
+ ########################################
36
+
37
+
38
+ if sys.version_info < (3, 8):
39
+ raise OSError(
40
+ f'Requires python (3, 8), got {sys.version_info} from {sys.executable}') # noqa
41
+
42
+
43
+ ########################################
44
+
45
+
46
+ # ../compat.py
47
+ T = ta.TypeVar('T')
48
+
49
+ # ../states.py
50
+ ProcessState = int # ta.TypeAlias
51
+ SupervisorState = int # ta.TypeAlias
52
+
53
+
54
+ ########################################
55
+ # ../compat.py
56
+
57
+
58
+ def as_bytes(s: ta.Union[str, bytes], encoding: str = 'utf8') -> bytes:
59
+ if isinstance(s, bytes):
60
+ return s
61
+ else:
62
+ return s.encode(encoding)
63
+
64
+
65
+ def as_string(s: ta.Union[str, bytes], encoding='utf8') -> str:
66
+ if isinstance(s, str):
67
+ return s
68
+ else:
69
+ return s.decode(encoding)
70
+
71
+
72
+ def compact_traceback() -> ta.Tuple[
73
+ ta.Tuple[str, str, int],
74
+ ta.Type[BaseException],
75
+ BaseException,
76
+ types.TracebackType,
77
+ ]:
78
+ t, v, tb = sys.exc_info()
79
+ tbinfo = []
80
+ if not tb:
81
+ raise RuntimeError('No traceback')
82
+ while tb:
83
+ tbinfo.append((
84
+ tb.tb_frame.f_code.co_filename,
85
+ tb.tb_frame.f_code.co_name,
86
+ str(tb.tb_lineno),
87
+ ))
88
+ tb = tb.tb_next
89
+
90
+ # just to be safe
91
+ del tb
92
+
93
+ file, function, line = tbinfo[-1]
94
+ info = ' '.join(['[%s|%s|%s]' % x for x in tbinfo]) # noqa
95
+ return (file, function, line), t, v, info # type: ignore
96
+
97
+
98
+ def find_prefix_at_end(haystack: bytes, needle: bytes) -> int:
99
+ l = len(needle) - 1
100
+ while l and not haystack.endswith(needle[:l]):
101
+ l -= 1
102
+ return l
103
+
104
+
105
+ class ExitNow(Exception): # noqa
106
+ pass
107
+
108
+
109
+ ##
110
+
111
+
112
+ def decode_wait_status(sts: int) -> ta.Tuple[int, str]:
113
+ """
114
+ Decode the status returned by wait() or waitpid().
115
+
116
+ Return a tuple (exitstatus, message) where exitstatus is the exit status, or -1 if the process was killed by a
117
+ signal; and message is a message telling what happened. It is the caller's responsibility to display the message.
118
+ """
119
+ if os.WIFEXITED(sts):
120
+ es = os.WEXITSTATUS(sts) & 0xffff
121
+ msg = f'exit status {es}'
122
+ return es, msg
123
+ elif os.WIFSIGNALED(sts):
124
+ sig = os.WTERMSIG(sts)
125
+ msg = f'terminated by {signame(sig)}'
126
+ if hasattr(os, 'WCOREDUMP'):
127
+ iscore = os.WCOREDUMP(sts)
128
+ else:
129
+ iscore = bool(sts & 0x80)
130
+ if iscore:
131
+ msg += ' (core dumped)'
132
+ return -1, msg
133
+ else:
134
+ msg = 'unknown termination cause 0x%04x' % sts # noqa
135
+ return -1, msg
136
+
137
+
138
+ _signames: ta.Optional[ta.Mapping[int, str]] = None
139
+
140
+
141
+ def signame(sig: int) -> str:
142
+ global _signames
143
+ if _signames is None:
144
+ _signames = _init_signames()
145
+ return _signames.get(sig) or 'signal %d' % sig
146
+
147
+
148
+ def _init_signames() -> ta.Dict[int, str]:
149
+ d = {}
150
+ for k, v in signal.__dict__.items():
151
+ k_startswith = getattr(k, 'startswith', None)
152
+ if k_startswith is None:
153
+ continue
154
+ if k_startswith('SIG') and not k_startswith('SIG_'):
155
+ d[v] = k
156
+ return d
157
+
158
+
159
+ class SignalReceiver:
160
+ def __init__(self) -> None:
161
+ super().__init__()
162
+ self._signals_recvd: ta.List[int] = []
163
+
164
+ def receive(self, sig: int, frame: ta.Any) -> None:
165
+ if sig not in self._signals_recvd:
166
+ self._signals_recvd.append(sig)
167
+
168
+ def install(self, *sigs: int) -> None:
169
+ for sig in sigs:
170
+ signal.signal(sig, self.receive)
171
+
172
+ def get_signal(self) -> ta.Optional[int]:
173
+ if self._signals_recvd:
174
+ sig = self._signals_recvd.pop(0)
175
+ else:
176
+ sig = None
177
+ return sig
178
+
179
+
180
+ def readfd(fd: int) -> bytes:
181
+ try:
182
+ data = os.read(fd, 2 << 16) # 128K
183
+ except OSError as why:
184
+ if why.args[0] not in (errno.EWOULDBLOCK, errno.EBADF, errno.EINTR):
185
+ raise
186
+ data = b''
187
+ return data
188
+
189
+
190
+ def try_unlink(path: str) -> bool:
191
+ try:
192
+ os.unlink(path)
193
+ except OSError:
194
+ return False
195
+ return True
196
+
197
+
198
+ def close_fd(fd: int) -> bool:
199
+ try:
200
+ os.close(fd)
201
+ except OSError:
202
+ return False
203
+ return True
204
+
205
+
206
+ def mktempfile(suffix: str, prefix: str, dir: str) -> str: # noqa
207
+ fd, filename = tempfile.mkstemp(suffix, prefix, dir)
208
+ os.close(fd)
209
+ return filename
210
+
211
+
212
+ def real_exit(code: int) -> None:
213
+ os._exit(code) # noqa
214
+
215
+
216
+ def get_path() -> ta.Sequence[str]:
217
+ """Return a list corresponding to $PATH, or a default."""
218
+ path = ['/bin', '/usr/bin', '/usr/local/bin']
219
+ if 'PATH' in os.environ:
220
+ p = os.environ['PATH']
221
+ if p:
222
+ path = p.split(os.pathsep)
223
+ return path
224
+
225
+
226
+ def normalize_path(v: str) -> str:
227
+ return os.path.normpath(os.path.abspath(os.path.expanduser(v)))
228
+
229
+
230
+ ANSI_ESCAPE_BEGIN = b'\x1b['
231
+ ANSI_TERMINATORS = (b'H', b'f', b'A', b'B', b'C', b'D', b'R', b's', b'u', b'J', b'K', b'h', b'l', b'p', b'm')
232
+
233
+
234
+ def strip_escapes(s):
235
+ """Remove all ANSI color escapes from the given string."""
236
+ result = b''
237
+ show = 1
238
+ i = 0
239
+ l = len(s)
240
+ while i < l:
241
+ if show == 0 and s[i:i + 1] in ANSI_TERMINATORS:
242
+ show = 1
243
+ elif show:
244
+ n = s.find(ANSI_ESCAPE_BEGIN, i)
245
+ if n == -1:
246
+ return result + s[i:]
247
+ else:
248
+ result = result + s[i:n]
249
+ i = n
250
+ show = 0
251
+ i += 1
252
+ return result
253
+
254
+
255
+ ########################################
256
+ # ../datatypes.py
257
+
258
+
259
+ class Automatic:
260
+ pass
261
+
262
+
263
+ class Syslog:
264
+ """TODO deprecated; remove this special 'syslog' filename in the future"""
265
+
266
+
267
+ LOGFILE_NONES = ('none', 'off', None)
268
+ LOGFILE_AUTOS = (Automatic, 'auto')
269
+ LOGFILE_SYSLOGS = (Syslog, 'syslog')
270
+
271
+
272
+ def logfile_name(val):
273
+ if hasattr(val, 'lower'):
274
+ coerced = val.lower()
275
+ else:
276
+ coerced = val
277
+
278
+ if coerced in LOGFILE_NONES:
279
+ return None
280
+ elif coerced in LOGFILE_AUTOS:
281
+ return Automatic
282
+ elif coerced in LOGFILE_SYSLOGS:
283
+ return Syslog
284
+ else:
285
+ return existing_dirpath(val)
286
+
287
+
288
+ def name_to_uid(name: str) -> int:
289
+ try:
290
+ uid = int(name)
291
+ except ValueError:
292
+ try:
293
+ pwdrec = pwd.getpwnam(name)
294
+ except KeyError:
295
+ raise ValueError(f'Invalid user name {name}') # noqa
296
+ uid = pwdrec[2]
297
+ else:
298
+ try:
299
+ pwd.getpwuid(uid) # check if uid is valid
300
+ except KeyError:
301
+ raise ValueError(f'Invalid user id {name}') # noqa
302
+ return uid
303
+
304
+
305
+ def name_to_gid(name: str) -> int:
306
+ try:
307
+ gid = int(name)
308
+ except ValueError:
309
+ try:
310
+ grprec = grp.getgrnam(name)
311
+ except KeyError:
312
+ raise ValueError(f'Invalid group name {name}') # noqa
313
+ gid = grprec[2]
314
+ else:
315
+ try:
316
+ grp.getgrgid(gid) # check if gid is valid
317
+ except KeyError:
318
+ raise ValueError(f'Invalid group id {name}') # noqa
319
+ return gid
320
+
321
+
322
+ def gid_for_uid(uid: int) -> int:
323
+ pwrec = pwd.getpwuid(uid)
324
+ return pwrec[3]
325
+
326
+
327
+ def octal_type(arg: ta.Union[str, int]) -> int:
328
+ if isinstance(arg, int):
329
+ return arg
330
+ try:
331
+ return int(arg, 8)
332
+ except (TypeError, ValueError):
333
+ raise ValueError(f'{arg} can not be converted to an octal type') # noqa
334
+
335
+
336
+ def existing_directory(v: str) -> str:
337
+ nv = os.path.expanduser(v)
338
+ if os.path.isdir(nv):
339
+ return nv
340
+ raise ValueError(f'{v} is not an existing directory')
341
+
342
+
343
+ def existing_dirpath(v: str) -> str:
344
+ nv = os.path.expanduser(v)
345
+ dir = os.path.dirname(nv) # noqa
346
+ if not dir:
347
+ # relative pathname with no directory component
348
+ return nv
349
+ if os.path.isdir(dir):
350
+ return nv
351
+ raise ValueError(f'The directory named as part of the path {v} does not exist')
352
+
353
+
354
+ def logging_level(value: ta.Union[str, int]) -> int:
355
+ if isinstance(value, int):
356
+ return value
357
+ s = str(value).lower()
358
+ level = logging.getLevelNamesMapping().get(s.upper())
359
+ if level is None:
360
+ raise ValueError(f'bad logging level name {value!r}')
361
+ return level
362
+
363
+
364
+ class SuffixMultiplier:
365
+ # d is a dictionary of suffixes to integer multipliers. If no suffixes match, default is the multiplier. Matches
366
+ # are case insensitive. Return values are in the fundamental unit.
367
+ def __init__(self, d, default=1):
368
+ super().__init__()
369
+ self._d = d
370
+ self._default = default
371
+ # all keys must be the same size
372
+ self._keysz = None
373
+ for k in d:
374
+ if self._keysz is None:
375
+ self._keysz = len(k)
376
+ elif self._keysz != len(k): # type: ignore
377
+ raise ValueError(k)
378
+
379
+ def __call__(self, v: ta.Union[str, int]) -> int:
380
+ if isinstance(v, int):
381
+ return v
382
+ v = v.lower()
383
+ for s, m in self._d.items():
384
+ if v[-self._keysz:] == s: # type: ignore
385
+ return int(v[:-self._keysz]) * m # type: ignore
386
+ return int(v) * self._default
387
+
388
+
389
+ byte_size = SuffixMultiplier({
390
+ 'kb': 1024,
391
+ 'mb': 1024 * 1024,
392
+ 'gb': 1024 * 1024 * 1024,
393
+ })
394
+
395
+
396
+ # all valid signal numbers
397
+ SIGNUMS = [getattr(signal, k) for k in dir(signal) if k.startswith('SIG')]
398
+
399
+
400
+ def signal_number(value: ta.Union[int, str]) -> int:
401
+ try:
402
+ num = int(value)
403
+ except (ValueError, TypeError):
404
+ name = value.strip().upper() # type: ignore
405
+ if not name.startswith('SIG'):
406
+ name = f'SIG{name}'
407
+ num = getattr(signal, name, None) # type: ignore
408
+ if num is None:
409
+ raise ValueError(f'value {value!r} is not a valid signal name') # noqa
410
+ if num not in SIGNUMS:
411
+ raise ValueError(f'value {value!r} is not a valid signal number')
412
+ return num
413
+
414
+
415
+ class RestartWhenExitUnexpected:
416
+ pass
417
+
418
+
419
+ class RestartUnconditionally:
420
+ pass
421
+
422
+
423
+ ########################################
424
+ # ../exceptions.py
425
+
426
+
427
+ class ProcessError(Exception):
428
+ """ Specialized exceptions used when attempting to start a process """
429
+
430
+
431
+ class BadCommandError(ProcessError):
432
+ """ Indicates the command could not be parsed properly. """
433
+
434
+
435
+ class NotExecutableError(ProcessError):
436
+ """ Indicates that the filespec cannot be executed because its path
437
+ resolves to a file which is not executable, or which is a directory. """
438
+
439
+
440
+ class NotFoundError(ProcessError):
441
+ """ Indicates that the filespec cannot be executed because it could not be found """
442
+
443
+
444
+ class NoPermissionError(ProcessError):
445
+ """
446
+ Indicates that the file cannot be executed because the supervisor process does not possess the appropriate UNIX
447
+ filesystem permission to execute the file.
448
+ """
449
+
450
+
451
+ ########################################
452
+ # ../poller.py
453
+
454
+
455
+ log = logging.getLogger(__name__)
456
+
457
+
458
+ class BasePoller(abc.ABC):
459
+
460
+ def __init__(self) -> None:
461
+ super().__init__()
462
+
463
+ @abc.abstractmethod
464
+ def register_readable(self, fd: int) -> None:
465
+ raise NotImplementedError
466
+
467
+ @abc.abstractmethod
468
+ def register_writable(self, fd: int) -> None:
469
+ raise NotImplementedError
470
+
471
+ @abc.abstractmethod
472
+ def unregister_readable(self, fd: int) -> None:
473
+ raise NotImplementedError
474
+
475
+ @abc.abstractmethod
476
+ def unregister_writable(self, fd: int) -> None:
477
+ raise NotImplementedError
478
+
479
+ @abc.abstractmethod
480
+ def poll(self, timeout: ta.Optional[float]) -> ta.Tuple[ta.List[int], ta.List[int]]:
481
+ raise NotImplementedError
482
+
483
+ def before_daemonize(self) -> None: # noqa
484
+ pass
485
+
486
+ def after_daemonize(self) -> None: # noqa
487
+ pass
488
+
489
+ def close(self) -> None: # noqa
490
+ pass
491
+
492
+
493
+ class SelectPoller(BasePoller):
494
+
495
+ def __init__(self) -> None:
496
+ super().__init__()
497
+
498
+ self._readables: ta.Set[int] = set()
499
+ self._writables: ta.Set[int] = set()
500
+
501
+ def register_readable(self, fd: int) -> None:
502
+ self._readables.add(fd)
503
+
504
+ def register_writable(self, fd: int) -> None:
505
+ self._writables.add(fd)
506
+
507
+ def unregister_readable(self, fd: int) -> None:
508
+ self._readables.discard(fd)
509
+
510
+ def unregister_writable(self, fd: int) -> None:
511
+ self._writables.discard(fd)
512
+
513
+ def unregister_all(self) -> None:
514
+ self._readables.clear()
515
+ self._writables.clear()
516
+
517
+ def poll(self, timeout: ta.Optional[float]) -> ta.Tuple[ta.List[int], ta.List[int]]:
518
+ try:
519
+ r, w, x = select.select(
520
+ self._readables,
521
+ self._writables,
522
+ [], timeout,
523
+ )
524
+ except OSError as err:
525
+ if err.args[0] == errno.EINTR:
526
+ log.debug('EINTR encountered in poll')
527
+ return [], []
528
+ if err.args[0] == errno.EBADF:
529
+ log.debug('EBADF encountered in poll')
530
+ self.unregister_all()
531
+ return [], []
532
+ raise
533
+ return r, w
534
+
535
+
536
+ class PollPoller(BasePoller):
537
+ _READ = select.POLLIN | select.POLLPRI | select.POLLHUP
538
+ _WRITE = select.POLLOUT
539
+
540
+ def __init__(self) -> None:
541
+ super().__init__()
542
+
543
+ self._poller = select.poll()
544
+ self._readables: set[int] = set()
545
+ self._writables: set[int] = set()
546
+
547
+ def register_readable(self, fd: int) -> None:
548
+ self._poller.register(fd, self._READ)
549
+ self._readables.add(fd)
550
+
551
+ def register_writable(self, fd: int) -> None:
552
+ self._poller.register(fd, self._WRITE)
553
+ self._writables.add(fd)
554
+
555
+ def unregister_readable(self, fd: int) -> None:
556
+ self._readables.discard(fd)
557
+ self._poller.unregister(fd)
558
+ if fd in self._writables:
559
+ self._poller.register(fd, self._WRITE)
560
+
561
+ def unregister_writable(self, fd: int) -> None:
562
+ self._writables.discard(fd)
563
+ self._poller.unregister(fd)
564
+ if fd in self._readables:
565
+ self._poller.register(fd, self._READ)
566
+
567
+ def poll(self, timeout: ta.Optional[float]) -> ta.Tuple[ta.List[int], ta.List[int]]:
568
+ fds = self._poll_fds(timeout) # type: ignore
569
+ readables, writables = [], []
570
+ for fd, eventmask in fds:
571
+ if self._ignore_invalid(fd, eventmask):
572
+ continue
573
+ if eventmask & self._READ:
574
+ readables.append(fd)
575
+ if eventmask & self._WRITE:
576
+ writables.append(fd)
577
+ return readables, writables
578
+
579
+ def _poll_fds(self, timeout: float) -> ta.List[ta.Tuple[int, int]]:
580
+ try:
581
+ return self._poller.poll(timeout * 1000)
582
+ except OSError as err:
583
+ if err.args[0] == errno.EINTR:
584
+ log.debug('EINTR encountered in poll')
585
+ return []
586
+ raise
587
+
588
+ def _ignore_invalid(self, fd: int, eventmask: int) -> bool:
589
+ if eventmask & select.POLLNVAL:
590
+ # POLLNVAL means `fd` value is invalid, not open. When a process quits it's `fd`s are closed so there is no
591
+ # more reason to keep this `fd` registered If the process restarts it's `fd`s are registered again.
592
+ self._poller.unregister(fd)
593
+ self._readables.discard(fd)
594
+ self._writables.discard(fd)
595
+ return True
596
+ return False
597
+
598
+
599
+ class KqueuePoller(BasePoller):
600
+ max_events = 1000
601
+
602
+ def __init__(self) -> None:
603
+ super().__init__()
604
+
605
+ self._kqueue: ta.Optional[ta.Any] = select.kqueue()
606
+ self._readables: set[int] = set()
607
+ self._writables: set[int] = set()
608
+
609
+ def register_readable(self, fd: int) -> None:
610
+ self._readables.add(fd)
611
+ kevent = select.kevent(fd, filter=select.KQ_FILTER_READ, flags=select.KQ_EV_ADD)
612
+ self._kqueue_control(fd, kevent)
613
+
614
+ def register_writable(self, fd: int) -> None:
615
+ self._writables.add(fd)
616
+ kevent = select.kevent(fd, filter=select.KQ_FILTER_WRITE, flags=select.KQ_EV_ADD)
617
+ self._kqueue_control(fd, kevent)
618
+
619
+ def unregister_readable(self, fd: int) -> None:
620
+ kevent = select.kevent(fd, filter=select.KQ_FILTER_READ, flags=select.KQ_EV_DELETE)
621
+ self._readables.discard(fd)
622
+ self._kqueue_control(fd, kevent)
623
+
624
+ def unregister_writable(self, fd: int) -> None:
625
+ kevent = select.kevent(fd, filter=select.KQ_FILTER_WRITE, flags=select.KQ_EV_DELETE)
626
+ self._writables.discard(fd)
627
+ self._kqueue_control(fd, kevent)
628
+
629
+ def _kqueue_control(self, fd: int, kevent: 'select.kevent') -> None:
630
+ try:
631
+ self._kqueue.control([kevent], 0) # type: ignore
632
+ except OSError as error:
633
+ if error.errno == errno.EBADF:
634
+ log.debug('EBADF encountered in kqueue. Invalid file descriptor %s', fd)
635
+ else:
636
+ raise
637
+
638
+ def poll(self, timeout: ta.Optional[float]) -> ta.Tuple[ta.List[int], ta.List[int]]:
639
+ readables, writables = [], [] # type: ignore
640
+
641
+ try:
642
+ kevents = self._kqueue.control(None, self.max_events, timeout) # type: ignore
643
+ except OSError as error:
644
+ if error.errno == errno.EINTR:
645
+ log.debug('EINTR encountered in poll')
646
+ return readables, writables
647
+ raise
648
+
649
+ for kevent in kevents:
650
+ if kevent.filter == select.KQ_FILTER_READ:
651
+ readables.append(kevent.ident)
652
+ if kevent.filter == select.KQ_FILTER_WRITE:
653
+ writables.append(kevent.ident)
654
+
655
+ return readables, writables
656
+
657
+ def before_daemonize(self) -> None:
658
+ self.close()
659
+
660
+ def after_daemonize(self) -> None:
661
+ self._kqueue = select.kqueue()
662
+ for fd in self._readables:
663
+ self.register_readable(fd)
664
+ for fd in self._writables:
665
+ self.register_writable(fd)
666
+
667
+ def close(self) -> None:
668
+ self._kqueue.close() # type: ignore
669
+ self._kqueue = None
670
+
671
+
672
+ Poller: ta.Type[BasePoller]
673
+ if hasattr(select, 'kqueue'):
674
+ Poller = KqueuePoller
675
+ elif hasattr(select, 'poll'):
676
+ Poller = PollPoller
677
+ else:
678
+ Poller = SelectPoller
679
+
680
+
681
+ ########################################
682
+ # ../../../omlish/lite/check.py
683
+
684
+
685
+ def check_isinstance(v: T, spec: ta.Union[ta.Type[T], tuple]) -> T:
686
+ if not isinstance(v, spec):
687
+ raise TypeError(v)
688
+ return v
689
+
690
+
691
+ def check_not_isinstance(v: T, spec: ta.Union[type, tuple]) -> T:
692
+ if isinstance(v, spec):
693
+ raise TypeError(v)
694
+ return v
695
+
696
+
697
+ def check_not_none(v: ta.Optional[T]) -> T:
698
+ if v is None:
699
+ raise ValueError
700
+ return v
701
+
702
+
703
+ def check_not(v: ta.Any) -> None:
704
+ if v:
705
+ raise ValueError(v)
706
+ return v
707
+
708
+
709
+ def check_non_empty_str(v: ta.Optional[str]) -> str:
710
+ if not v:
711
+ raise ValueError
712
+ return v
713
+
714
+
715
+ def check_state(v: bool, msg: str = 'Illegal state') -> None:
716
+ if not v:
717
+ raise ValueError(msg)
718
+
719
+
720
+ def check_equal(l: T, r: T) -> T:
721
+ if l != r:
722
+ raise ValueError(l, r)
723
+ return l
724
+
725
+
726
+ def check_not_equal(l: T, r: T) -> T:
727
+ if l == r:
728
+ raise ValueError(l, r)
729
+ return l
730
+
731
+
732
+ def check_single(vs: ta.Iterable[T]) -> T:
733
+ [v] = vs
734
+ return v
735
+
736
+
737
+ ########################################
738
+ # ../../../omlish/lite/json.py
739
+
740
+
741
+ ##
742
+
743
+
744
+ JSON_PRETTY_INDENT = 2
745
+
746
+ JSON_PRETTY_KWARGS: ta.Mapping[str, ta.Any] = dict(
747
+ indent=JSON_PRETTY_INDENT,
748
+ )
749
+
750
+ json_dump_pretty: ta.Callable[..., bytes] = functools.partial(json.dump, **JSON_PRETTY_KWARGS) # type: ignore
751
+ json_dumps_pretty: ta.Callable[..., str] = functools.partial(json.dumps, **JSON_PRETTY_KWARGS)
752
+
753
+
754
+ ##
755
+
756
+
757
+ JSON_COMPACT_SEPARATORS = (',', ':')
758
+
759
+ JSON_COMPACT_KWARGS: ta.Mapping[str, ta.Any] = dict(
760
+ indent=None,
761
+ separators=JSON_COMPACT_SEPARATORS,
762
+ )
763
+
764
+ json_dump_compact: ta.Callable[..., bytes] = functools.partial(json.dump, **JSON_COMPACT_KWARGS) # type: ignore
765
+ json_dumps_compact: ta.Callable[..., str] = functools.partial(json.dumps, **JSON_COMPACT_KWARGS)
766
+
767
+
768
+ ########################################
769
+ # ../configs.py
770
+
771
+
772
+ @dc.dataclass(frozen=True)
773
+ class ServerConfig:
774
+ user: ta.Optional[str] = None
775
+ nodaemon: bool = False
776
+ umask: int = 0o22
777
+ directory: ta.Optional[str] = None
778
+ logfile: str = 'supervisord.log'
779
+ logfile_maxbytes: int = 50 * 1024 * 1024
780
+ logfile_backups: int = 10
781
+ loglevel: int = logging.INFO
782
+ pidfile: str = 'supervisord.pid'
783
+ identifier: str = 'supervisor'
784
+ child_logdir: str = '/dev/null'
785
+ minfds: int = 1024
786
+ minprocs: int = 200
787
+ nocleanup: bool = False
788
+ strip_ansi: bool = False
789
+ silent: bool = False
790
+
791
+ groups: ta.Optional[ta.Sequence['ProcessGroupConfig']] = None
792
+
793
+ @classmethod
794
+ def new(
795
+ cls,
796
+ umask: ta.Union[int, str] = 0o22,
797
+ directory: ta.Optional[str] = None,
798
+ logfile: str = 'supervisord.log',
799
+ logfile_maxbytes: ta.Union[int, str] = 50 * 1024 * 1024,
800
+ loglevel: ta.Union[int, str] = logging.INFO,
801
+ pidfile: str = 'supervisord.pid',
802
+ child_logdir: ta.Optional[str] = None,
803
+ **kwargs: ta.Any,
804
+ ) -> 'ServerConfig':
805
+ return cls(
806
+ umask=octal_type(umask),
807
+ directory=existing_directory(directory) if directory is not None else None,
808
+ logfile=existing_dirpath(logfile),
809
+ logfile_maxbytes=byte_size(logfile_maxbytes),
810
+ loglevel=logging_level(loglevel),
811
+ pidfile=existing_dirpath(pidfile),
812
+ child_logdir=child_logdir if child_logdir else tempfile.gettempdir(),
813
+ **kwargs,
814
+ )
815
+
816
+
817
+ @dc.dataclass(frozen=True)
818
+ class ProcessGroupConfig:
819
+ name: str
820
+
821
+ priority: int = 999
822
+
823
+ processes: ta.Optional[ta.Sequence['ProcessConfig']] = None
824
+
825
+
826
+ @dc.dataclass(frozen=True)
827
+ class ProcessConfig:
828
+ name: str
829
+ command: str
830
+
831
+ uid: ta.Optional[int] = None
832
+ directory: ta.Optional[str] = None
833
+ umask: ta.Optional[int] = None
834
+ priority: int = 999
835
+
836
+ autostart: bool = True
837
+ autorestart: str = 'unexpected'
838
+
839
+ startsecs: int = 1
840
+ startretries: int = 3
841
+
842
+ numprocs: int = 1
843
+ numprocs_start: int = 0
844
+
845
+ @dc.dataclass(frozen=True)
846
+ class Log:
847
+ file: ta.Optional[str] = None
848
+ capture_maxbytes: ta.Optional[int] = None
849
+ events_enabled: bool = False
850
+ syslog: bool = False
851
+ backups: ta.Optional[int] = None
852
+ maxbytes: ta.Optional[int] = None
853
+
854
+ stdout: Log = Log()
855
+ stderr: Log = Log()
856
+
857
+ stopsignal: int = signal.SIGTERM
858
+ stopwaitsecs: int = 10
859
+ stopasgroup: bool = False
860
+
861
+ killasgroup: bool = False
862
+
863
+ exitcodes: ta.Iterable[int] = (0,)
864
+
865
+ redirect_stderr: bool = False
866
+
867
+ environment: ta.Optional[ta.Mapping[str, str]] = None
868
+
869
+
870
+ ########################################
871
+ # ../states.py
872
+
873
+
874
+ ##
875
+
876
+
877
+ def _names_by_code(states: ta.Any) -> ta.Dict[int, str]:
878
+ d = {}
879
+ for name in states.__dict__:
880
+ if not name.startswith('__'):
881
+ code = getattr(states, name)
882
+ d[code] = name
883
+ return d
884
+
885
+
886
+ ##
887
+
888
+
889
+ class ProcessStates:
890
+ STOPPED = 0
891
+ STARTING = 10
892
+ RUNNING = 20
893
+ BACKOFF = 30
894
+ STOPPING = 40
895
+ EXITED = 100
896
+ FATAL = 200
897
+ UNKNOWN = 1000
898
+
899
+
900
+ STOPPED_STATES = (
901
+ ProcessStates.STOPPED,
902
+ ProcessStates.EXITED,
903
+ ProcessStates.FATAL,
904
+ ProcessStates.UNKNOWN,
905
+ )
906
+
907
+ RUNNING_STATES = (
908
+ ProcessStates.RUNNING,
909
+ ProcessStates.BACKOFF,
910
+ ProcessStates.STARTING,
911
+ )
912
+
913
+ SIGNALLABLE_STATES = (
914
+ ProcessStates.RUNNING,
915
+ ProcessStates.STARTING,
916
+ ProcessStates.STOPPING,
917
+ )
918
+
919
+
920
+ _process_states_by_code = _names_by_code(ProcessStates)
921
+
922
+
923
+ def get_process_state_description(code: ProcessState) -> str:
924
+ return check_not_none(_process_states_by_code.get(code))
925
+
926
+
927
+ ##
928
+
929
+
930
+ class SupervisorStates:
931
+ FATAL = 2
932
+ RUNNING = 1
933
+ RESTARTING = 0
934
+ SHUTDOWN = -1
935
+
936
+
937
+ _supervisor_states_by_code = _names_by_code(SupervisorStates)
938
+
939
+
940
+ def get_supervisor_state_description(code: SupervisorState) -> str:
941
+ return check_not_none(_supervisor_states_by_code.get(code))
942
+
943
+
944
+ ########################################
945
+ # ../../../omlish/lite/logs.py
946
+ """
947
+ TODO:
948
+ - translate json keys
949
+ - debug
950
+ """
951
+
952
+
953
+ log = logging.getLogger(__name__)
954
+
955
+
956
+ ##
957
+
958
+
959
+ class TidLogFilter(logging.Filter):
960
+
961
+ def filter(self, record):
962
+ record.tid = threading.get_native_id()
963
+ return True
964
+
965
+
966
+ ##
967
+
968
+
969
+ class JsonLogFormatter(logging.Formatter):
970
+
971
+ KEYS: ta.Mapping[str, bool] = {
972
+ 'name': False,
973
+ 'msg': False,
974
+ 'args': False,
975
+ 'levelname': False,
976
+ 'levelno': False,
977
+ 'pathname': False,
978
+ 'filename': False,
979
+ 'module': False,
980
+ 'exc_info': True,
981
+ 'exc_text': True,
982
+ 'stack_info': True,
983
+ 'lineno': False,
984
+ 'funcName': False,
985
+ 'created': False,
986
+ 'msecs': False,
987
+ 'relativeCreated': False,
988
+ 'thread': False,
989
+ 'threadName': False,
990
+ 'processName': False,
991
+ 'process': False,
992
+ }
993
+
994
+ def format(self, record: logging.LogRecord) -> str:
995
+ dct = {
996
+ k: v
997
+ for k, o in self.KEYS.items()
998
+ for v in [getattr(record, k)]
999
+ if not (o and v is None)
1000
+ }
1001
+ return json_dumps_compact(dct)
1002
+
1003
+
1004
+ ##
1005
+
1006
+
1007
+ STANDARD_LOG_FORMAT_PARTS = [
1008
+ ('asctime', '%(asctime)-15s'),
1009
+ ('process', 'pid=%(process)-6s'),
1010
+ ('thread', 'tid=%(thread)x'),
1011
+ ('levelname', '%(levelname)s'),
1012
+ ('name', '%(name)s'),
1013
+ ('separator', '::'),
1014
+ ('message', '%(message)s'),
1015
+ ]
1016
+
1017
+
1018
+ class StandardLogFormatter(logging.Formatter):
1019
+
1020
+ @staticmethod
1021
+ def build_log_format(parts: ta.Iterable[ta.Tuple[str, str]]) -> str:
1022
+ return ' '.join(v for k, v in parts)
1023
+
1024
+ converter = datetime.datetime.fromtimestamp # type: ignore
1025
+
1026
+ def formatTime(self, record, datefmt=None):
1027
+ ct = self.converter(record.created) # type: ignore
1028
+ if datefmt:
1029
+ return ct.strftime(datefmt) # noqa
1030
+ else:
1031
+ t = ct.strftime("%Y-%m-%d %H:%M:%S") # noqa
1032
+ return '%s.%03d' % (t, record.msecs)
1033
+
1034
+
1035
+ ##
1036
+
1037
+
1038
+ class ProxyLogFilterer(logging.Filterer):
1039
+ def __init__(self, underlying: logging.Filterer) -> None: # noqa
1040
+ self._underlying = underlying
1041
+
1042
+ @property
1043
+ def underlying(self) -> logging.Filterer:
1044
+ return self._underlying
1045
+
1046
+ @property
1047
+ def filters(self):
1048
+ return self._underlying.filters
1049
+
1050
+ @filters.setter
1051
+ def filters(self, filters):
1052
+ self._underlying.filters = filters
1053
+
1054
+ def addFilter(self, filter): # noqa
1055
+ self._underlying.addFilter(filter)
1056
+
1057
+ def removeFilter(self, filter): # noqa
1058
+ self._underlying.removeFilter(filter)
1059
+
1060
+ def filter(self, record):
1061
+ return self._underlying.filter(record)
1062
+
1063
+
1064
+ class ProxyLogHandler(ProxyLogFilterer, logging.Handler):
1065
+ def __init__(self, underlying: logging.Handler) -> None: # noqa
1066
+ ProxyLogFilterer.__init__(self, underlying)
1067
+
1068
+ _underlying: logging.Handler
1069
+
1070
+ @property
1071
+ def underlying(self) -> logging.Handler:
1072
+ return self._underlying
1073
+
1074
+ def get_name(self):
1075
+ return self._underlying.get_name()
1076
+
1077
+ def set_name(self, name):
1078
+ self._underlying.set_name(name)
1079
+
1080
+ @property
1081
+ def name(self):
1082
+ return self._underlying.name
1083
+
1084
+ @property
1085
+ def level(self):
1086
+ return self._underlying.level
1087
+
1088
+ @level.setter
1089
+ def level(self, level):
1090
+ self._underlying.level = level
1091
+
1092
+ @property
1093
+ def formatter(self):
1094
+ return self._underlying.formatter
1095
+
1096
+ @formatter.setter
1097
+ def formatter(self, formatter):
1098
+ self._underlying.formatter = formatter
1099
+
1100
+ def createLock(self):
1101
+ self._underlying.createLock()
1102
+
1103
+ def acquire(self):
1104
+ self._underlying.acquire()
1105
+
1106
+ def release(self):
1107
+ self._underlying.release()
1108
+
1109
+ def setLevel(self, level):
1110
+ self._underlying.setLevel(level)
1111
+
1112
+ def format(self, record):
1113
+ return self._underlying.format(record)
1114
+
1115
+ def emit(self, record):
1116
+ self._underlying.emit(record)
1117
+
1118
+ def handle(self, record):
1119
+ return self._underlying.handle(record)
1120
+
1121
+ def setFormatter(self, fmt):
1122
+ self._underlying.setFormatter(fmt)
1123
+
1124
+ def flush(self):
1125
+ self._underlying.flush()
1126
+
1127
+ def close(self):
1128
+ self._underlying.close()
1129
+
1130
+ def handleError(self, record):
1131
+ self._underlying.handleError(record)
1132
+
1133
+
1134
+ ##
1135
+
1136
+
1137
+ class StandardLogHandler(ProxyLogHandler):
1138
+ pass
1139
+
1140
+
1141
+ ##
1142
+
1143
+
1144
+ @contextlib.contextmanager
1145
+ def _locking_logging_module_lock() -> ta.Iterator[None]:
1146
+ if hasattr(logging, '_acquireLock'):
1147
+ logging._acquireLock() # noqa
1148
+ try:
1149
+ yield
1150
+ finally:
1151
+ logging._releaseLock() # type: ignore # noqa
1152
+
1153
+ elif hasattr(logging, '_lock'):
1154
+ # https://github.com/python/cpython/commit/74723e11109a320e628898817ab449b3dad9ee96
1155
+ with logging._lock: # noqa
1156
+ yield
1157
+
1158
+ else:
1159
+ raise Exception("Can't find lock in logging module")
1160
+
1161
+
1162
+ def configure_standard_logging(
1163
+ level: ta.Union[int, str] = logging.INFO,
1164
+ *,
1165
+ json: bool = False,
1166
+ target: ta.Optional[logging.Logger] = None,
1167
+ force: bool = False,
1168
+ ) -> ta.Optional[StandardLogHandler]:
1169
+ with _locking_logging_module_lock():
1170
+ if target is None:
1171
+ target = logging.root
1172
+
1173
+ #
1174
+
1175
+ if not force:
1176
+ if any(isinstance(h, StandardLogHandler) for h in list(target.handlers)):
1177
+ return None
1178
+
1179
+ #
1180
+
1181
+ handler = logging.StreamHandler()
1182
+
1183
+ #
1184
+
1185
+ formatter: logging.Formatter
1186
+ if json:
1187
+ formatter = JsonLogFormatter()
1188
+ else:
1189
+ formatter = StandardLogFormatter(StandardLogFormatter.build_log_format(STANDARD_LOG_FORMAT_PARTS))
1190
+ handler.setFormatter(formatter)
1191
+
1192
+ #
1193
+
1194
+ handler.addFilter(TidLogFilter())
1195
+
1196
+ #
1197
+
1198
+ target.addHandler(handler)
1199
+
1200
+ #
1201
+
1202
+ if level is not None:
1203
+ target.setLevel(level)
1204
+
1205
+ #
1206
+
1207
+ return StandardLogHandler(handler)
1208
+
1209
+
1210
+ ########################################
1211
+ # ../events.py
1212
+
1213
+
1214
+ class EventCallbacks:
1215
+ def __init__(self) -> None:
1216
+ super().__init__()
1217
+
1218
+ self._callbacks: ta.List[ta.Tuple[type, ta.Callable]] = []
1219
+
1220
+ def subscribe(self, type, callback): # noqa
1221
+ self._callbacks.append((type, callback))
1222
+
1223
+ def unsubscribe(self, type, callback): # noqa
1224
+ self._callbacks.remove((type, callback))
1225
+
1226
+ def notify(self, event):
1227
+ for type, callback in self._callbacks: # noqa
1228
+ if isinstance(event, type):
1229
+ callback(event)
1230
+
1231
+ def clear(self):
1232
+ self._callbacks[:] = []
1233
+
1234
+
1235
+ EVENT_CALLBACKS = EventCallbacks()
1236
+
1237
+ notify_event = EVENT_CALLBACKS.notify
1238
+ clear_events = EVENT_CALLBACKS.clear
1239
+
1240
+
1241
+ class Event:
1242
+ """Abstract event type """
1243
+
1244
+
1245
+ class ProcessLogEvent(Event):
1246
+ """Abstract"""
1247
+ channel: ta.Optional[str] = None
1248
+
1249
+ def __init__(self, process, pid, data):
1250
+ super().__init__()
1251
+ self.process = process
1252
+ self.pid = pid
1253
+ self.data = data
1254
+
1255
+ def payload(self):
1256
+ groupname = ''
1257
+ if self.process.group is not None:
1258
+ groupname = self.process.group.config.name
1259
+ try:
1260
+ data = as_string(self.data)
1261
+ except UnicodeDecodeError:
1262
+ data = f'Undecodable: {self.data!r}'
1263
+ fmt = as_string('processname:%s groupname:%s pid:%s channel:%s\n%s')
1264
+ result = fmt % (
1265
+ as_string(self.process.config.name),
1266
+ as_string(groupname),
1267
+ self.pid,
1268
+ as_string(self.channel), # type: ignore
1269
+ data,
1270
+ )
1271
+ return result
1272
+
1273
+
1274
+ class ProcessLogStdoutEvent(ProcessLogEvent):
1275
+ channel = 'stdout'
1276
+
1277
+
1278
+ class ProcessLogStderrEvent(ProcessLogEvent):
1279
+ channel = 'stderr'
1280
+
1281
+
1282
+ class ProcessCommunicationEvent(Event):
1283
+ """ Abstract """
1284
+ # event mode tokens
1285
+ BEGIN_TOKEN = b'<!--XSUPERVISOR:BEGIN-->'
1286
+ END_TOKEN = b'<!--XSUPERVISOR:END-->'
1287
+
1288
+ def __init__(self, process, pid, data):
1289
+ super().__init__()
1290
+ self.process = process
1291
+ self.pid = pid
1292
+ self.data = data
1293
+
1294
+ def payload(self):
1295
+ groupname = ''
1296
+ if self.process.group is not None:
1297
+ groupname = self.process.group.config.name
1298
+ try:
1299
+ data = as_string(self.data)
1300
+ except UnicodeDecodeError:
1301
+ data = f'Undecodable: {self.data!r}'
1302
+ return f'processname:{self.process.config.name} groupname:{groupname} pid:{self.pid}\n{data}'
1303
+
1304
+
1305
+ class ProcessCommunicationStdoutEvent(ProcessCommunicationEvent):
1306
+ channel = 'stdout'
1307
+
1308
+
1309
+ class ProcessCommunicationStderrEvent(ProcessCommunicationEvent):
1310
+ channel = 'stderr'
1311
+
1312
+
1313
+ class RemoteCommunicationEvent(Event):
1314
+ def __init__(self, type, data): # noqa
1315
+ super().__init__()
1316
+ self.type = type
1317
+ self.data = data
1318
+
1319
+ def payload(self):
1320
+ return f'type:{self.type}\n{self.data}'
1321
+
1322
+
1323
+ class SupervisorStateChangeEvent(Event):
1324
+ """ Abstract class """
1325
+
1326
+ def payload(self):
1327
+ return ''
1328
+
1329
+
1330
+ class SupervisorRunningEvent(SupervisorStateChangeEvent):
1331
+ pass
1332
+
1333
+
1334
+ class SupervisorStoppingEvent(SupervisorStateChangeEvent):
1335
+ pass
1336
+
1337
+
1338
+ class EventRejectedEvent: # purposely does not subclass Event
1339
+ def __init__(self, process, event):
1340
+ super().__init__()
1341
+ self.process = process
1342
+ self.event = event
1343
+
1344
+
1345
+ class ProcessStateEvent(Event):
1346
+ """ Abstract class, never raised directly """
1347
+ frm = None
1348
+ to = None
1349
+
1350
+ def __init__(self, process, from_state, expected=True):
1351
+ super().__init__()
1352
+ self.process = process
1353
+ self.from_state = from_state
1354
+ self.expected = expected
1355
+ # we eagerly render these so if the process pid, etc changes beneath
1356
+ # us, we stash the values at the time the event was sent
1357
+ self.extra_values = self.get_extra_values()
1358
+
1359
+ def payload(self):
1360
+ groupname = ''
1361
+ if self.process.group is not None:
1362
+ groupname = self.process.group.config.name
1363
+ l = [
1364
+ ('processname', self.process.config.name),
1365
+ ('groupname', groupname),
1366
+ ('from_state', get_process_state_description(self.from_state)),
1367
+ ]
1368
+ l.extend(self.extra_values)
1369
+ s = ' '.join([f'{name}:{val}' for name, val in l])
1370
+ return s
1371
+
1372
+ def get_extra_values(self):
1373
+ return []
1374
+
1375
+
1376
+ class ProcessStateFatalEvent(ProcessStateEvent):
1377
+ pass
1378
+
1379
+
1380
+ class ProcessStateUnknownEvent(ProcessStateEvent):
1381
+ pass
1382
+
1383
+
1384
+ class ProcessStateStartingOrBackoffEvent(ProcessStateEvent):
1385
+ def get_extra_values(self):
1386
+ return [('tries', int(self.process.backoff))]
1387
+
1388
+
1389
+ class ProcessStateBackoffEvent(ProcessStateStartingOrBackoffEvent):
1390
+ pass
1391
+
1392
+
1393
+ class ProcessStateStartingEvent(ProcessStateStartingOrBackoffEvent):
1394
+ pass
1395
+
1396
+
1397
+ class ProcessStateExitedEvent(ProcessStateEvent):
1398
+ def get_extra_values(self):
1399
+ return [('expected', int(self.expected)), ('pid', self.process.pid)]
1400
+
1401
+
1402
+ class ProcessStateRunningEvent(ProcessStateEvent):
1403
+ def get_extra_values(self):
1404
+ return [('pid', self.process.pid)]
1405
+
1406
+
1407
+ class ProcessStateStoppingEvent(ProcessStateEvent):
1408
+ def get_extra_values(self):
1409
+ return [('pid', self.process.pid)]
1410
+
1411
+
1412
+ class ProcessStateStoppedEvent(ProcessStateEvent):
1413
+ def get_extra_values(self):
1414
+ return [('pid', self.process.pid)]
1415
+
1416
+
1417
+ class ProcessGroupEvent(Event):
1418
+ def __init__(self, group):
1419
+ super().__init__()
1420
+ self.group = group
1421
+
1422
+ def payload(self):
1423
+ return f'groupname:{self.group}\n'
1424
+
1425
+
1426
+ class ProcessGroupAddedEvent(ProcessGroupEvent):
1427
+ pass
1428
+
1429
+
1430
+ class ProcessGroupRemovedEvent(ProcessGroupEvent):
1431
+ pass
1432
+
1433
+
1434
+ class TickEvent(Event):
1435
+ """ Abstract """
1436
+
1437
+ def __init__(self, when, supervisord):
1438
+ super().__init__()
1439
+ self.when = when
1440
+ self.supervisord = supervisord
1441
+
1442
+ def payload(self):
1443
+ return f'when:{self.when}'
1444
+
1445
+
1446
+ class Tick5Event(TickEvent):
1447
+ period = 5
1448
+
1449
+
1450
+ class Tick60Event(TickEvent):
1451
+ period = 60
1452
+
1453
+
1454
+ class Tick3600Event(TickEvent):
1455
+ period = 3600
1456
+
1457
+
1458
+ TICK_EVENTS = [ # imported elsewhere
1459
+ Tick5Event,
1460
+ Tick60Event,
1461
+ Tick3600Event,
1462
+ ]
1463
+
1464
+
1465
+ class EventTypes:
1466
+ EVENT = Event # abstract
1467
+
1468
+ PROCESS_STATE = ProcessStateEvent # abstract
1469
+ PROCESS_STATE_STOPPED = ProcessStateStoppedEvent
1470
+ PROCESS_STATE_EXITED = ProcessStateExitedEvent
1471
+ PROCESS_STATE_STARTING = ProcessStateStartingEvent
1472
+ PROCESS_STATE_STOPPING = ProcessStateStoppingEvent
1473
+ PROCESS_STATE_BACKOFF = ProcessStateBackoffEvent
1474
+ PROCESS_STATE_FATAL = ProcessStateFatalEvent
1475
+ PROCESS_STATE_RUNNING = ProcessStateRunningEvent
1476
+ PROCESS_STATE_UNKNOWN = ProcessStateUnknownEvent
1477
+
1478
+ PROCESS_COMMUNICATION = ProcessCommunicationEvent # abstract
1479
+ PROCESS_COMMUNICATION_STDOUT = ProcessCommunicationStdoutEvent
1480
+ PROCESS_COMMUNICATION_STDERR = ProcessCommunicationStderrEvent
1481
+
1482
+ PROCESS_LOG = ProcessLogEvent
1483
+ PROCESS_LOG_STDOUT = ProcessLogStdoutEvent
1484
+ PROCESS_LOG_STDERR = ProcessLogStderrEvent
1485
+
1486
+ REMOTE_COMMUNICATION = RemoteCommunicationEvent
1487
+
1488
+ SUPERVISOR_STATE_CHANGE = SupervisorStateChangeEvent # abstract
1489
+ SUPERVISOR_STATE_CHANGE_RUNNING = SupervisorRunningEvent
1490
+ SUPERVISOR_STATE_CHANGE_STOPPING = SupervisorStoppingEvent
1491
+
1492
+ TICK = TickEvent # abstract
1493
+ TICK_5 = Tick5Event
1494
+ TICK_60 = Tick60Event
1495
+ TICK_3600 = Tick3600Event
1496
+
1497
+ PROCESS_GROUP = ProcessGroupEvent # abstract
1498
+ PROCESS_GROUP_ADDED = ProcessGroupAddedEvent
1499
+ PROCESS_GROUP_REMOVED = ProcessGroupRemovedEvent
1500
+
1501
+
1502
+ def get_event_name_by_type(requested):
1503
+ for name, typ in EventTypes.__dict__.items():
1504
+ if typ is requested:
1505
+ return name
1506
+ return None
1507
+
1508
+
1509
+ def register(name, event):
1510
+ setattr(EventTypes, name, event)
1511
+
1512
+
1513
+ ########################################
1514
+ # ../types.py
1515
+
1516
+
1517
+ class AbstractServerContext(abc.ABC):
1518
+ @property
1519
+ @abc.abstractmethod
1520
+ def config(self) -> ServerConfig:
1521
+ raise NotImplementedError
1522
+
1523
+ @property
1524
+ @abc.abstractmethod
1525
+ def state(self) -> SupervisorState:
1526
+ raise NotImplementedError
1527
+
1528
+ @abc.abstractmethod
1529
+ def set_state(self, state: SupervisorState) -> None:
1530
+ raise NotImplementedError
1531
+
1532
+ @property
1533
+ @abc.abstractmethod
1534
+ def pid_history(self) -> ta.Dict[int, 'AbstractSubprocess']:
1535
+ raise NotImplementedError
1536
+
1537
+
1538
+ class AbstractSubprocess(abc.ABC):
1539
+ @property
1540
+ @abc.abstractmethod
1541
+ def pid(self) -> int:
1542
+ raise NotImplementedError
1543
+
1544
+ @property
1545
+ @abc.abstractmethod
1546
+ def config(self) -> ProcessConfig:
1547
+ raise NotImplementedError
1548
+
1549
+ @property
1550
+ @abc.abstractmethod
1551
+ def context(self) -> AbstractServerContext:
1552
+ raise NotImplementedError
1553
+
1554
+ @abc.abstractmethod
1555
+ def finish(self, sts: int) -> None:
1556
+ raise NotImplementedError
1557
+
1558
+
1559
+ ########################################
1560
+ # ../context.py
1561
+
1562
+
1563
+ log = logging.getLogger(__name__)
1564
+
1565
+
1566
+ class ServerContext(AbstractServerContext):
1567
+ first = False
1568
+ test = False
1569
+
1570
+ ##
1571
+
1572
+ def __init__(self, config: ServerConfig) -> None:
1573
+ super().__init__()
1574
+
1575
+ self._config = config
1576
+
1577
+ self._pid_history: ta.Dict[int, AbstractSubprocess] = {}
1578
+ self._state: SupervisorState = SupervisorStates.RUNNING
1579
+
1580
+ self.signal_receiver = SignalReceiver()
1581
+
1582
+ self.poller = Poller()
1583
+
1584
+ if self.config.user is not None:
1585
+ uid = name_to_uid(self.config.user)
1586
+ self.uid = uid
1587
+ self.gid = gid_for_uid(uid)
1588
+ else:
1589
+ self.uid = None
1590
+ self.gid = None
1591
+
1592
+ self.unlink_pidfile = False
1593
+
1594
+ @property
1595
+ def config(self) -> ServerConfig:
1596
+ return self._config
1597
+
1598
+ @property
1599
+ def state(self) -> SupervisorState:
1600
+ return self._state
1601
+
1602
+ def set_state(self, state: SupervisorState) -> None:
1603
+ self._state = state
1604
+
1605
+ @property
1606
+ def pid_history(self) -> ta.Dict[int, AbstractSubprocess]:
1607
+ return self._pid_history
1608
+
1609
+ uid: ta.Optional[int]
1610
+ gid: ta.Optional[int]
1611
+
1612
+ ##
1613
+
1614
+ def set_signals(self) -> None:
1615
+ self.signal_receiver.install(
1616
+ signal.SIGTERM,
1617
+ signal.SIGINT,
1618
+ signal.SIGQUIT,
1619
+ signal.SIGHUP,
1620
+ signal.SIGCHLD,
1621
+ signal.SIGUSR2,
1622
+ )
1623
+
1624
+ def waitpid(self) -> ta.Tuple[ta.Optional[int], ta.Optional[int]]:
1625
+ # Need pthread_sigmask here to avoid concurrent sigchld, but Python doesn't offer in Python < 3.4. There is
1626
+ # still a race condition here; we can get a sigchld while we're sitting in the waitpid call. However, AFAICT, if
1627
+ # waitpid is interrupted by SIGCHLD, as long as we call waitpid again (which happens every so often during the
1628
+ # normal course in the mainloop), we'll eventually reap the child that we tried to reap during the interrupted
1629
+ # call. At least on Linux, this appears to be true, or at least stopping 50 processes at once never left zombies
1630
+ # lying around.
1631
+ try:
1632
+ pid, sts = os.waitpid(-1, os.WNOHANG)
1633
+ except OSError as exc:
1634
+ code = exc.args[0]
1635
+ if code not in (errno.ECHILD, errno.EINTR):
1636
+ log.critical('waitpid error %r; a process may not be cleaned up properly', code)
1637
+ if code == errno.EINTR:
1638
+ log.debug('EINTR during reap')
1639
+ pid, sts = None, None
1640
+ return pid, sts
1641
+
1642
+ def set_uid_or_exit(self) -> None:
1643
+ """
1644
+ Set the uid of the supervisord process. Called during supervisord startup only. No return value. Exits the
1645
+ process via usage() if privileges could not be dropped.
1646
+ """
1647
+ if self.uid is None:
1648
+ if os.getuid() == 0:
1649
+ warnings.warn(
1650
+ 'Supervisor is running as root. Privileges were not dropped because no user is specified in the '
1651
+ 'config file. If you intend to run as root, you can set user=root in the config file to avoid '
1652
+ 'this message.',
1653
+ )
1654
+ else:
1655
+ msg = drop_privileges(self.uid)
1656
+ if msg is None:
1657
+ log.info('Set uid to user %s succeeded', self.uid)
1658
+ else: # failed to drop privileges
1659
+ raise RuntimeError(msg)
1660
+
1661
+ def set_rlimits_or_exit(self) -> None:
1662
+ """
1663
+ Set the rlimits of the supervisord process. Called during supervisord startup only. No return value. Exits
1664
+ the process via usage() if any rlimits could not be set.
1665
+ """
1666
+
1667
+ limits = []
1668
+
1669
+ if hasattr(resource, 'RLIMIT_NOFILE'):
1670
+ limits.append({
1671
+ 'msg': (
1672
+ 'The minimum number of file descriptors required to run this process is %(min_limit)s as per the '
1673
+ '"minfds" command-line argument or config file setting. The current environment will only allow '
1674
+ 'you to open %(hard)s file descriptors. Either raise the number of usable file descriptors in '
1675
+ 'your environment (see README.rst) or lower the minfds setting in the config file to allow the '
1676
+ 'process to start.'
1677
+ ),
1678
+ 'min': self.config.minfds,
1679
+ 'resource': resource.RLIMIT_NOFILE,
1680
+ 'name': 'RLIMIT_NOFILE',
1681
+ })
1682
+
1683
+ if hasattr(resource, 'RLIMIT_NPROC'):
1684
+ limits.append({
1685
+ 'msg': (
1686
+ 'The minimum number of available processes required to run this program is %(min_limit)s as per '
1687
+ 'the "minprocs" command-line argument or config file setting. The current environment will only '
1688
+ 'allow you to open %(hard)s processes. Either raise the number of usable processes in your '
1689
+ 'environment (see README.rst) or lower the minprocs setting in the config file to allow the '
1690
+ 'program to start.'
1691
+ ),
1692
+ 'min': self.config.minprocs,
1693
+ 'resource': resource.RLIMIT_NPROC,
1694
+ 'name': 'RLIMIT_NPROC',
1695
+ })
1696
+
1697
+ for limit in limits:
1698
+ min_limit = limit['min']
1699
+ res = limit['resource']
1700
+ msg = limit['msg']
1701
+ name = limit['name']
1702
+
1703
+ soft, hard = resource.getrlimit(res) # type: ignore
1704
+
1705
+ # -1 means unlimited
1706
+ if soft < min_limit and soft != -1: # type: ignore
1707
+ if hard < min_limit and hard != -1: # type: ignore
1708
+ # setrlimit should increase the hard limit if we are root, if not then setrlimit raises and we print
1709
+ # usage
1710
+ hard = min_limit # type: ignore
1711
+
1712
+ try:
1713
+ resource.setrlimit(res, (min_limit, hard)) # type: ignore
1714
+ log.info('Increased %s limit to %s', name, min_limit)
1715
+ except (resource.error, ValueError):
1716
+ raise RuntimeError(msg % dict( # type: ignore # noqa
1717
+ min_limit=min_limit,
1718
+ res=res,
1719
+ name=name,
1720
+ soft=soft,
1721
+ hard=hard,
1722
+ ))
1723
+
1724
+ def cleanup(self) -> None:
1725
+ if self.unlink_pidfile:
1726
+ try_unlink(self.config.pidfile)
1727
+ self.poller.close()
1728
+
1729
+ def cleanup_fds(self) -> None:
1730
+ # try to close any leaked file descriptors (for reload)
1731
+ start = 5
1732
+ os.closerange(start, self.config.minfds)
1733
+
1734
+ def clear_auto_child_logdir(self) -> None:
1735
+ # must be called after realize()
1736
+ child_logdir = self.config.child_logdir
1737
+ fnre = re.compile(rf'.+?---{self.config.identifier}-\S+\.log\.?\d{{0,4}}')
1738
+ try:
1739
+ filenames = os.listdir(child_logdir)
1740
+ except OSError:
1741
+ log.warning('Could not clear child_log dir')
1742
+ return
1743
+
1744
+ for filename in filenames:
1745
+ if fnre.match(filename):
1746
+ pathname = os.path.join(child_logdir, filename)
1747
+ try:
1748
+ os.remove(pathname)
1749
+ except OSError:
1750
+ log.warning('Failed to clean up %r', pathname)
1751
+
1752
+ def daemonize(self) -> None:
1753
+ self.poller.before_daemonize()
1754
+ self._daemonize()
1755
+ self.poller.after_daemonize()
1756
+
1757
+ def _daemonize(self) -> None:
1758
+ # To daemonize, we need to become the leader of our own session (process) group. If we do not, signals sent to
1759
+ # our parent process will also be sent to us. This might be bad because signals such as SIGINT can be sent to
1760
+ # our parent process during normal (uninteresting) operations such as when we press Ctrl-C in the parent
1761
+ # terminal window to escape from a logtail command. To disassociate ourselves from our parent's session group we
1762
+ # use os.setsid. It means "set session id", which has the effect of disassociating a process from is current
1763
+ # session and process group and setting itself up as a new session leader.
1764
+ #
1765
+ # Unfortunately we cannot call setsid if we're already a session group leader, so we use "fork" to make a copy
1766
+ # of ourselves that is guaranteed to not be a session group leader.
1767
+ #
1768
+ # We also change directories, set stderr and stdout to null, and change our umask.
1769
+ #
1770
+ # This explanation was (gratefully) garnered from
1771
+ # http://www.cems.uwe.ac.uk/~irjohnso/coursenotes/lrc/system/daemons/d3.htm
1772
+
1773
+ pid = os.fork()
1774
+ if pid != 0:
1775
+ # Parent
1776
+ log.debug('supervisord forked; parent exiting')
1777
+ real_exit(0)
1778
+ # Child
1779
+ log.info('daemonizing the supervisord process')
1780
+ if self.config.directory:
1781
+ try:
1782
+ os.chdir(self.config.directory)
1783
+ except OSError as err:
1784
+ log.critical("can't chdir into %r: %s", self.config.directory, err)
1785
+ else:
1786
+ log.info('set current directory: %r', self.config.directory)
1787
+ os.dup2(0, os.open('/dev/null', os.O_RDONLY))
1788
+ os.dup2(1, os.open('/dev/null', os.O_WRONLY))
1789
+ os.dup2(2, os.open('/dev/null', os.O_WRONLY))
1790
+ os.setsid()
1791
+ os.umask(self.config.umask)
1792
+ # XXX Stevens, in his Advanced Unix book, section 13.3 (page 417) recommends calling umask(0) and closing unused
1793
+ # file descriptors. In his Network Programming book, he additionally recommends ignoring SIGHUP and forking
1794
+ # again after the setsid() call, for obscure SVR4 reasons.
1795
+
1796
+ def get_auto_child_log_name(self, name: str, identifier: str, channel: str) -> str:
1797
+ prefix = f'{name}-{channel}---{identifier}-'
1798
+ logfile = mktempfile(
1799
+ suffix='.log',
1800
+ prefix=prefix,
1801
+ dir=self.config.child_logdir,
1802
+ )
1803
+ return logfile
1804
+
1805
+ def get_signal(self) -> ta.Optional[int]:
1806
+ return self.signal_receiver.get_signal()
1807
+
1808
+ def write_pidfile(self) -> None:
1809
+ pid = os.getpid()
1810
+ try:
1811
+ with open(self.config.pidfile, 'w') as f:
1812
+ f.write(f'{pid}\n')
1813
+ except OSError:
1814
+ log.critical('could not write pidfile %s', self.config.pidfile)
1815
+ else:
1816
+ self.unlink_pidfile = True
1817
+ log.info('supervisord started with pid %s', pid)
1818
+
1819
+
1820
+ def drop_privileges(user: ta.Union[int, str, None]) -> ta.Optional[str]:
1821
+ """
1822
+ Drop privileges to become the specified user, which may be a username or uid. Called for supervisord startup
1823
+ and when spawning subprocesses. Returns None on success or a string error message if privileges could not be
1824
+ dropped.
1825
+ """
1826
+ if user is None:
1827
+ return 'No user specified to setuid to!'
1828
+
1829
+ # get uid for user, which can be a number or username
1830
+ try:
1831
+ uid = int(user)
1832
+ except ValueError:
1833
+ try:
1834
+ pwrec = pwd.getpwnam(user) # type: ignore
1835
+ except KeyError:
1836
+ return f"Can't find username {user!r}"
1837
+ uid = pwrec[2]
1838
+ else:
1839
+ try:
1840
+ pwrec = pwd.getpwuid(uid)
1841
+ except KeyError:
1842
+ return f"Can't find uid {uid!r}"
1843
+
1844
+ current_uid = os.getuid()
1845
+
1846
+ if current_uid == uid:
1847
+ # do nothing and return successfully if the uid is already the current one. this allows a supervisord
1848
+ # running as an unprivileged user "foo" to start a process where the config has "user=foo" (same user) in
1849
+ # it.
1850
+ return None
1851
+
1852
+ if current_uid != 0:
1853
+ return "Can't drop privilege as nonroot user"
1854
+
1855
+ gid = pwrec[3]
1856
+ if hasattr(os, 'setgroups'):
1857
+ user = pwrec[0]
1858
+ groups = [grprec[2] for grprec in grp.getgrall() if user in grprec[3]]
1859
+
1860
+ # always put our primary gid first in this list, otherwise we can lose group info since sometimes the first
1861
+ # group in the setgroups list gets overwritten on the subsequent setgid call (at least on freebsd 9 with
1862
+ # python 2.7 - this will be safe though for all unix /python version combos)
1863
+ groups.insert(0, gid)
1864
+ try:
1865
+ os.setgroups(groups)
1866
+ except OSError:
1867
+ return 'Could not set groups of effective user'
1868
+ try:
1869
+ os.setgid(gid)
1870
+ except OSError:
1871
+ return 'Could not set group id of effective user'
1872
+ os.setuid(uid)
1873
+ return None
1874
+
1875
+
1876
+ def make_pipes(stderr=True) -> ta.Mapping[str, int]:
1877
+ """
1878
+ Create pipes for parent to child stdin/stdout/stderr communications. Open fd in non-blocking mode so we can
1879
+ read them in the mainloop without blocking. If stderr is False, don't create a pipe for stderr.
1880
+ """
1881
+
1882
+ pipes: ta.Dict[str, ta.Optional[int]] = {
1883
+ 'child_stdin': None,
1884
+ 'stdin': None,
1885
+ 'stdout': None,
1886
+ 'child_stdout': None,
1887
+ 'stderr': None,
1888
+ 'child_stderr': None,
1889
+ }
1890
+ try:
1891
+ stdin, child_stdin = os.pipe()
1892
+ pipes['child_stdin'], pipes['stdin'] = stdin, child_stdin
1893
+ stdout, child_stdout = os.pipe()
1894
+ pipes['stdout'], pipes['child_stdout'] = stdout, child_stdout
1895
+ if stderr:
1896
+ stderr, child_stderr = os.pipe()
1897
+ pipes['stderr'], pipes['child_stderr'] = stderr, child_stderr
1898
+ for fd in (pipes['stdout'], pipes['stderr'], pipes['stdin']):
1899
+ if fd is not None:
1900
+ flags = fcntl.fcntl(fd, fcntl.F_GETFL) | os.O_NDELAY
1901
+ fcntl.fcntl(fd, fcntl.F_SETFL, flags)
1902
+ return pipes # type: ignore
1903
+ except OSError:
1904
+ for fd in pipes.values():
1905
+ if fd is not None:
1906
+ close_fd(fd)
1907
+ raise
1908
+
1909
+
1910
+ def close_parent_pipes(pipes: ta.Mapping[str, int]) -> None:
1911
+ for fdname in ('stdin', 'stdout', 'stderr'):
1912
+ fd = pipes.get(fdname)
1913
+ if fd is not None:
1914
+ close_fd(fd)
1915
+
1916
+
1917
+ def close_child_pipes(pipes: ta.Mapping[str, int]) -> None:
1918
+ for fdname in ('child_stdin', 'child_stdout', 'child_stderr'):
1919
+ fd = pipes.get(fdname)
1920
+ if fd is not None:
1921
+ close_fd(fd)
1922
+
1923
+
1924
+ def check_execv_args(filename, argv, st) -> None:
1925
+ if st is None:
1926
+ raise NotFoundError(f"can't find command {filename!r}")
1927
+
1928
+ elif stat.S_ISDIR(st[stat.ST_MODE]):
1929
+ raise NotExecutableError(f'command at {filename!r} is a directory')
1930
+
1931
+ elif not (stat.S_IMODE(st[stat.ST_MODE]) & 0o111):
1932
+ raise NotExecutableError(f'command at {filename!r} is not executable')
1933
+
1934
+ elif not os.access(filename, os.X_OK):
1935
+ raise NoPermissionError(f'no permission to run command {filename!r}')
1936
+
1937
+
1938
+ ########################################
1939
+ # ../dispatchers.py
1940
+
1941
+
1942
+ log = logging.getLogger(__name__)
1943
+
1944
+
1945
+ class Dispatcher(abc.ABC):
1946
+
1947
+ def __init__(self, process: AbstractSubprocess, channel: str, fd: int) -> None:
1948
+ super().__init__()
1949
+
1950
+ self._process = process # process which "owns" this dispatcher
1951
+ self._channel = channel # 'stderr' or 'stdout'
1952
+ self._fd = fd
1953
+ self._closed = False # True if close() has been called
1954
+
1955
+ def __repr__(self) -> str:
1956
+ return f'<{self.__class__.__name__} at {id(self)} for {self._process} ({self._channel})>'
1957
+
1958
+ @property
1959
+ def process(self) -> AbstractSubprocess:
1960
+ return self._process
1961
+
1962
+ @property
1963
+ def channel(self) -> str:
1964
+ return self._channel
1965
+
1966
+ @property
1967
+ def fd(self) -> int:
1968
+ return self._fd
1969
+
1970
+ @property
1971
+ def closed(self) -> bool:
1972
+ return self._closed
1973
+
1974
+ @abc.abstractmethod
1975
+ def readable(self) -> bool:
1976
+ raise NotImplementedError
1977
+
1978
+ @abc.abstractmethod
1979
+ def writable(self) -> bool:
1980
+ raise NotImplementedError
1981
+
1982
+ def handle_read_event(self) -> None:
1983
+ raise TypeError
1984
+
1985
+ def handle_write_event(self) -> None:
1986
+ raise TypeError
1987
+
1988
+ def handle_error(self) -> None:
1989
+ nil, t, v, tbinfo = compact_traceback()
1990
+
1991
+ log.critical('uncaptured python exception, closing channel %s (%s:%s %s)', repr(self), t, v, tbinfo)
1992
+ self.close()
1993
+
1994
+ def close(self) -> None:
1995
+ if not self._closed:
1996
+ log.debug('fd %s closed, stopped monitoring %s', self._fd, self)
1997
+ self._closed = True
1998
+
1999
+ def flush(self) -> None: # noqa
2000
+ pass
2001
+
2002
+
2003
+ class OutputDispatcher(Dispatcher):
2004
+ """
2005
+ Dispatcher for one channel (stdout or stderr) of one process. Serves several purposes:
2006
+
2007
+ - capture output sent within <!--XSUPERVISOR:BEGIN--> and <!--XSUPERVISOR:END--> tags and signal a
2008
+ ProcessCommunicationEvent by calling notify_event(event).
2009
+ - route the output to the appropriate log handlers as specified in the config.
2010
+ """
2011
+
2012
+ def __init__(self, process: AbstractSubprocess, event_type, fd):
2013
+ """
2014
+ Initialize the dispatcher.
2015
+
2016
+ `event_type` should be one of ProcessLogStdoutEvent or ProcessLogStderrEvent
2017
+ """
2018
+ super().__init__(process, event_type.channel, fd)
2019
+ self.event_type = event_type
2020
+
2021
+ self.lc: ProcessConfig.Log = getattr(process.config, self._channel)
2022
+
2023
+ self._init_normal_log()
2024
+ self._init_capture_log()
2025
+
2026
+ self._child_log = self._normal_log
2027
+
2028
+ self._capture_mode = False # are we capturing process event data
2029
+ self._output_buffer = b'' # data waiting to be logged
2030
+
2031
+ # all code below is purely for minor speedups
2032
+ begin_token = self.event_type.BEGIN_TOKEN
2033
+ end_token = self.event_type.END_TOKEN
2034
+ self.begin_token_data = (begin_token, len(begin_token))
2035
+ self.end_token_data = (end_token, len(end_token))
2036
+ self.main_log_level = logging.DEBUG
2037
+ config = self._process.config
2038
+ self.log_to_main_log = process.context.config.loglevel <= self.main_log_level
2039
+ self.stdout_events_enabled = config.stdout.events_enabled
2040
+ self.stderr_events_enabled = config.stderr.events_enabled
2041
+
2042
+ _child_log: ta.Optional[logging.Logger] # the current logger (normal_log or capture_log)
2043
+ _normal_log: ta.Optional[logging.Logger] # the "normal" (non-capture) logger
2044
+ _capture_log: ta.Optional[logging.Logger] # the logger used while we're in capture_mode
2045
+
2046
+ def _init_normal_log(self) -> None:
2047
+ """
2048
+ Configure the "normal" (non-capture) log for this channel of this process. Sets self.normal_log if logging is
2049
+ enabled.
2050
+ """
2051
+ config = self._process.config # noqa
2052
+ channel = self._channel # noqa
2053
+
2054
+ logfile = self.lc.file
2055
+ maxbytes = self.lc.maxbytes # noqa
2056
+ backups = self.lc.backups # noqa
2057
+ to_syslog = self.lc.syslog
2058
+
2059
+ if logfile or to_syslog:
2060
+ self._normal_log = logging.getLogger(__name__)
2061
+
2062
+ # if logfile:
2063
+ # loggers.handle_file(
2064
+ # self.normal_log,
2065
+ # filename=logfile,
2066
+ # fmt='%(message)s',
2067
+ # rotating=bool(maxbytes), # optimization
2068
+ # maxbytes=maxbytes,
2069
+ # backups=backups,
2070
+ # )
2071
+ #
2072
+ # if to_syslog:
2073
+ # loggers.handle_syslog(
2074
+ # self.normal_log,
2075
+ # fmt=config.name + ' %(message)s',
2076
+ # )
2077
+
2078
+ def _init_capture_log(self) -> None:
2079
+ """
2080
+ Configure the capture log for this process. This log is used to temporarily capture output when special output
2081
+ is detected. Sets self.capture_log if capturing is enabled.
2082
+ """
2083
+ capture_maxbytes = self.lc.capture_maxbytes
2084
+ if capture_maxbytes:
2085
+ self._capture_log = logging.getLogger(__name__)
2086
+ # loggers.handle_boundIO(
2087
+ # self._capture_log,
2088
+ # fmt='%(message)s',
2089
+ # maxbytes=capture_maxbytes,
2090
+ # )
2091
+
2092
+ def remove_logs(self):
2093
+ for log in (self._normal_log, self._capture_log):
2094
+ if log is not None:
2095
+ for handler in log.handlers:
2096
+ handler.remove() # type: ignore
2097
+ handler.reopen() # type: ignore
2098
+
2099
+ def reopen_logs(self):
2100
+ for log in (self._normal_log, self._capture_log):
2101
+ if log is not None:
2102
+ for handler in log.handlers:
2103
+ handler.reopen() # type: ignore
2104
+
2105
+ def _log(self, data):
2106
+ if data:
2107
+ if self._process.context.config.strip_ansi:
2108
+ data = strip_escapes(data)
2109
+ if self._child_log:
2110
+ self._child_log.info(data)
2111
+ if self.log_to_main_log:
2112
+ if not isinstance(data, bytes):
2113
+ text = data
2114
+ else:
2115
+ try:
2116
+ text = data.decode('utf-8')
2117
+ except UnicodeDecodeError:
2118
+ text = f'Undecodable: {data!r}'
2119
+ log.log(self.main_log_level, '%r %s output:\n%s', self._process.config.name, self._channel, text) # noqa
2120
+ if self._channel == 'stdout':
2121
+ if self.stdout_events_enabled:
2122
+ notify_event(ProcessLogStdoutEvent(self._process, self._process.pid, data))
2123
+ elif self.stderr_events_enabled:
2124
+ notify_event(ProcessLogStderrEvent(self._process, self._process.pid, data))
2125
+
2126
+ def record_output(self):
2127
+ if self._capture_log is None:
2128
+ # shortcut trying to find capture data
2129
+ data = self._output_buffer
2130
+ self._output_buffer = b''
2131
+ self._log(data)
2132
+ return
2133
+
2134
+ if self._capture_mode:
2135
+ token, tokenlen = self.end_token_data
2136
+ else:
2137
+ token, tokenlen = self.begin_token_data
2138
+
2139
+ if len(self._output_buffer) <= tokenlen:
2140
+ return # not enough data
2141
+
2142
+ data = self._output_buffer
2143
+ self._output_buffer = b''
2144
+
2145
+ try:
2146
+ before, after = data.split(token, 1)
2147
+ except ValueError:
2148
+ after = None
2149
+ index = find_prefix_at_end(data, token)
2150
+ if index:
2151
+ self._output_buffer = self._output_buffer + data[-index:]
2152
+ data = data[:-index]
2153
+ self._log(data)
2154
+ else:
2155
+ self._log(before)
2156
+ self.toggle_capture_mode()
2157
+ self._output_buffer = after # type: ignore
2158
+
2159
+ if after:
2160
+ self.record_output()
2161
+
2162
+ def toggle_capture_mode(self):
2163
+ self._capture_mode = not self._capture_mode
2164
+
2165
+ if self._capture_log is not None:
2166
+ if self._capture_mode:
2167
+ self._child_log = self._capture_log
2168
+ else:
2169
+ for handler in self._capture_log.handlers:
2170
+ handler.flush()
2171
+ data = self._capture_log.getvalue() # type: ignore
2172
+ channel = self._channel
2173
+ procname = self._process.config.name
2174
+ event = self.event_type(self._process, self._process.pid, data)
2175
+ notify_event(event)
2176
+
2177
+ log.debug('%r %s emitted a comm event', procname, channel)
2178
+ for handler in self._capture_log.handlers:
2179
+ handler.remove() # type: ignore
2180
+ handler.reopen() # type: ignore
2181
+ self._child_log = self._normal_log
2182
+
2183
+ def writable(self) -> bool:
2184
+ return False
2185
+
2186
+ def readable(self) -> bool:
2187
+ if self._closed:
2188
+ return False
2189
+ return True
2190
+
2191
+ def handle_read_event(self) -> None:
2192
+ data = readfd(self._fd)
2193
+ self._output_buffer += data
2194
+ self.record_output()
2195
+ if not data:
2196
+ # if we get no data back from the pipe, it means that the child process has ended. See
2197
+ # mail.python.org/pipermail/python-dev/2004-August/046850.html
2198
+ self.close()
2199
+
2200
+
2201
+ class InputDispatcher(Dispatcher):
2202
+
2203
+ def __init__(self, process: AbstractSubprocess, channel: str, fd: int) -> None:
2204
+ super().__init__(process, channel, fd)
2205
+ self._input_buffer = b''
2206
+
2207
+ def writable(self) -> bool:
2208
+ if self._input_buffer and not self._closed:
2209
+ return True
2210
+ return False
2211
+
2212
+ def readable(self) -> bool:
2213
+ return False
2214
+
2215
+ def flush(self) -> None:
2216
+ # other code depends on this raising EPIPE if the pipe is closed
2217
+ sent = os.write(self._fd, as_bytes(self._input_buffer))
2218
+ self._input_buffer = self._input_buffer[sent:]
2219
+
2220
+ def handle_write_event(self) -> None:
2221
+ if self._input_buffer:
2222
+ try:
2223
+ self.flush()
2224
+ except OSError as why:
2225
+ if why.args[0] == errno.EPIPE:
2226
+ self._input_buffer = b''
2227
+ self.close()
2228
+ else:
2229
+ raise
2230
+
2231
+
2232
+ ########################################
2233
+ # ../process.py
2234
+
2235
+
2236
+ log = logging.getLogger(__name__)
2237
+
2238
+
2239
+ @functools.total_ordering
2240
+ class Subprocess(AbstractSubprocess):
2241
+ """A class to manage a subprocess."""
2242
+
2243
+ # Initial state; overridden by instance variables
2244
+
2245
+ # pid = 0 # Subprocess pid; 0 when not running
2246
+ # config = None # ProcessConfig instance
2247
+ # state = None # process state code
2248
+ listener_state = None # listener state code (if we're an event listener)
2249
+ event = None # event currently being processed (if we're an event listener)
2250
+ laststart = 0. # Last time the subprocess was started; 0 if never
2251
+ laststop = 0. # Last time the subprocess was stopped; 0 if never
2252
+ last_stop_report = 0. # Last time "waiting for x to stop" logged, to throttle
2253
+ delay = 0. # If nonzero, delay starting or killing until this time
2254
+ administrative_stop = False # true if process has been stopped by an admin
2255
+ system_stop = False # true if process has been stopped by the system
2256
+ killing = False # true if we are trying to kill this process
2257
+ backoff = 0 # backoff counter (to startretries)
2258
+ dispatchers = None # asyncore output dispatchers (keyed by fd)
2259
+ pipes = None # map of channel name to file descriptor #
2260
+ exitstatus = None # status attached to dead process by finish()
2261
+ spawn_err = None # error message attached by spawn() if any
2262
+ group = None # ProcessGroup instance if process is in the group
2263
+
2264
+ def __init__(self, config: ProcessConfig, group: 'ProcessGroup', context: AbstractServerContext) -> None:
2265
+ super().__init__()
2266
+ self._config = config
2267
+ self.group = group
2268
+ self._context = context
2269
+ self._dispatchers: dict = {}
2270
+ self._pipes: dict = {}
2271
+ self.state = ProcessStates.STOPPED
2272
+ self._pid = 0
2273
+
2274
+ @property
2275
+ def pid(self) -> int:
2276
+ return self._pid
2277
+
2278
+ @property
2279
+ def config(self) -> ProcessConfig:
2280
+ return self._config
2281
+
2282
+ @property
2283
+ def context(self) -> AbstractServerContext:
2284
+ return self._context
2285
+
2286
+ def remove_logs(self) -> None:
2287
+ for dispatcher in self._dispatchers.values():
2288
+ if hasattr(dispatcher, 'remove_logs'):
2289
+ dispatcher.remove_logs()
2290
+
2291
+ def reopen_logs(self) -> None:
2292
+ for dispatcher in self._dispatchers.values():
2293
+ if hasattr(dispatcher, 'reopen_logs'):
2294
+ dispatcher.reopen_logs()
2295
+
2296
+ def drain(self) -> None:
2297
+ for dispatcher in self._dispatchers.values():
2298
+ # note that we *must* call readable() for every dispatcher, as it may have side effects for a given
2299
+ # dispatcher (eg. call handle_listener_state_change for event listener processes)
2300
+ if dispatcher.readable():
2301
+ dispatcher.handle_read_event()
2302
+ if dispatcher.writable():
2303
+ dispatcher.handle_write_event()
2304
+
2305
+ def write(self, chars: ta.Union[bytes, str]) -> None:
2306
+ if not self.pid or self.killing:
2307
+ raise OSError(errno.EPIPE, 'Process already closed')
2308
+
2309
+ stdin_fd = self._pipes['stdin']
2310
+ if stdin_fd is None:
2311
+ raise OSError(errno.EPIPE, 'Process has no stdin channel')
2312
+
2313
+ dispatcher = self._dispatchers[stdin_fd]
2314
+ if dispatcher.closed:
2315
+ raise OSError(errno.EPIPE, "Process' stdin channel is closed")
2316
+
2317
+ dispatcher.input_buffer += chars
2318
+ dispatcher.flush() # this must raise EPIPE if the pipe is closed
2319
+
2320
+ def _get_execv_args(self) -> ta.Tuple[str, ta.Sequence[str]]:
2321
+ """
2322
+ Internal: turn a program name into a file name, using $PATH, make sure it exists / is executable, raising a
2323
+ ProcessError if not
2324
+ """
2325
+ try:
2326
+ commandargs = shlex.split(self.config.command)
2327
+ except ValueError as e:
2328
+ raise BadCommandError(f"can't parse command {self.config.command!r}: {e}") # noqa
2329
+
2330
+ if commandargs:
2331
+ program = commandargs[0]
2332
+ else:
2333
+ raise BadCommandError('command is empty')
2334
+
2335
+ if '/' in program:
2336
+ filename = program
2337
+ try:
2338
+ st = os.stat(filename)
2339
+ except OSError:
2340
+ st = None
2341
+
2342
+ else:
2343
+ path = get_path()
2344
+ found = None
2345
+ st = None
2346
+ for dir in path: # noqa
2347
+ found = os.path.join(dir, program)
2348
+ try:
2349
+ st = os.stat(found)
2350
+ except OSError:
2351
+ pass
2352
+ else:
2353
+ break
2354
+ if st is None:
2355
+ filename = program
2356
+ else:
2357
+ filename = found # type: ignore
2358
+
2359
+ # check_execv_args will raise a ProcessError if the execv args are bogus, we break it out into a separate
2360
+ # options method call here only to service unit tests
2361
+ check_execv_args(filename, commandargs, st)
2362
+
2363
+ return filename, commandargs
2364
+
2365
+ event_map: ta.ClassVar[ta.Mapping[int, ta.Type[ProcessStateEvent]]] = {
2366
+ ProcessStates.BACKOFF: ProcessStateBackoffEvent,
2367
+ ProcessStates.FATAL: ProcessStateFatalEvent,
2368
+ ProcessStates.UNKNOWN: ProcessStateUnknownEvent,
2369
+ ProcessStates.STOPPED: ProcessStateStoppedEvent,
2370
+ ProcessStates.EXITED: ProcessStateExitedEvent,
2371
+ ProcessStates.RUNNING: ProcessStateRunningEvent,
2372
+ ProcessStates.STARTING: ProcessStateStartingEvent,
2373
+ ProcessStates.STOPPING: ProcessStateStoppingEvent,
2374
+ }
2375
+
2376
+ def change_state(self, new_state: ProcessState, expected: bool = True) -> bool:
2377
+ old_state = self.state
2378
+ if new_state is old_state:
2379
+ return False
2380
+
2381
+ self.state = new_state
2382
+ if new_state == ProcessStates.BACKOFF:
2383
+ now = time.time()
2384
+ self.backoff += 1
2385
+ self.delay = now + self.backoff
2386
+
2387
+ event_class = self.event_map.get(new_state)
2388
+ if event_class is not None:
2389
+ event = event_class(self, old_state, expected)
2390
+ notify_event(event)
2391
+
2392
+ return True
2393
+
2394
+ def _check_in_state(self, *states: ProcessState) -> None:
2395
+ if self.state not in states:
2396
+ current_state = get_process_state_description(self.state)
2397
+ allowable_states = ' '.join(map(get_process_state_description, states))
2398
+ processname = as_string(self.config.name)
2399
+ raise AssertionError('Assertion failed for %s: %s not in %s' % (processname, current_state, allowable_states)) # noqa
2400
+
2401
+ def _record_spawn_err(self, msg: str) -> None:
2402
+ self.spawn_err = msg
2403
+ log.info('spawn_err: %s', msg)
2404
+
2405
+ def spawn(self) -> ta.Optional[int]:
2406
+ processname = as_string(self.config.name)
2407
+
2408
+ if self.pid:
2409
+ log.warning('process \'%s\' already running', processname)
2410
+ return None
2411
+
2412
+ self.killing = False
2413
+ self.spawn_err = None
2414
+ self.exitstatus = None
2415
+ self.system_stop = False
2416
+ self.administrative_stop = False
2417
+
2418
+ self.laststart = time.time()
2419
+
2420
+ self._check_in_state(
2421
+ ProcessStates.EXITED,
2422
+ ProcessStates.FATAL,
2423
+ ProcessStates.BACKOFF,
2424
+ ProcessStates.STOPPED,
2425
+ )
2426
+
2427
+ self.change_state(ProcessStates.STARTING)
2428
+
2429
+ try:
2430
+ filename, argv = self._get_execv_args()
2431
+ except ProcessError as what:
2432
+ self._record_spawn_err(what.args[0])
2433
+ self._check_in_state(ProcessStates.STARTING)
2434
+ self.change_state(ProcessStates.BACKOFF)
2435
+ return None
2436
+
2437
+ try:
2438
+ self._dispatchers, self._pipes = self._make_dispatchers() # type: ignore
2439
+ except OSError as why:
2440
+ code = why.args[0]
2441
+ if code == errno.EMFILE:
2442
+ # too many file descriptors open
2443
+ msg = f"too many open files to spawn '{processname}'"
2444
+ else:
2445
+ msg = f"unknown error making dispatchers for '{processname}': {errno.errorcode.get(code, code)}"
2446
+ self._record_spawn_err(msg)
2447
+ self._check_in_state(ProcessStates.STARTING)
2448
+ self.change_state(ProcessStates.BACKOFF)
2449
+ return None
2450
+
2451
+ try:
2452
+ pid = os.fork()
2453
+ except OSError as why:
2454
+ code = why.args[0]
2455
+ if code == errno.EAGAIN:
2456
+ # process table full
2457
+ msg = f'Too many processes in process table to spawn \'{processname}\''
2458
+ else:
2459
+ msg = f'unknown error during fork for \'{processname}\': {errno.errorcode.get(code, code)}'
2460
+ self._record_spawn_err(msg)
2461
+ self._check_in_state(ProcessStates.STARTING)
2462
+ self.change_state(ProcessStates.BACKOFF)
2463
+ close_parent_pipes(self._pipes)
2464
+ close_child_pipes(self._pipes)
2465
+ return None
2466
+
2467
+ if pid != 0:
2468
+ return self._spawn_as_parent(pid)
2469
+
2470
+ else:
2471
+ self._spawn_as_child(filename, argv)
2472
+ return None
2473
+
2474
+ def _make_dispatchers(self) -> ta.Tuple[ta.Mapping[int, Dispatcher], ta.Mapping[str, int]]:
2475
+ use_stderr = not self.config.redirect_stderr
2476
+ p = make_pipes(use_stderr)
2477
+ stdout_fd, stderr_fd, stdin_fd = p['stdout'], p['stderr'], p['stdin']
2478
+ dispatchers: ta.Dict[int, Dispatcher] = {}
2479
+ etype: ta.Type[ProcessCommunicationEvent]
2480
+ if stdout_fd is not None:
2481
+ etype = ProcessCommunicationStdoutEvent
2482
+ dispatchers[stdout_fd] = OutputDispatcher(self, etype, stdout_fd)
2483
+ if stderr_fd is not None:
2484
+ etype = ProcessCommunicationStderrEvent
2485
+ dispatchers[stderr_fd] = OutputDispatcher(self, etype, stderr_fd)
2486
+ if stdin_fd is not None:
2487
+ dispatchers[stdin_fd] = InputDispatcher(self, 'stdin', stdin_fd)
2488
+ return dispatchers, p
2489
+
2490
+ def _spawn_as_parent(self, pid: int) -> int:
2491
+ # Parent
2492
+ self._pid = pid
2493
+ close_child_pipes(self._pipes)
2494
+ log.info('spawned: \'%s\' with pid %s', as_string(self.config.name), pid)
2495
+ self.spawn_err = None
2496
+ self.delay = time.time() + self.config.startsecs
2497
+ self.context.pid_history[pid] = self
2498
+ return pid
2499
+
2500
+ def _prepare_child_fds(self) -> None:
2501
+ os.dup2(self._pipes['child_stdin'], 0)
2502
+ os.dup2(self._pipes['child_stdout'], 1)
2503
+ if self.config.redirect_stderr:
2504
+ os.dup2(self._pipes['child_stdout'], 2)
2505
+ else:
2506
+ os.dup2(self._pipes['child_stderr'], 2)
2507
+ for i in range(3, self.context.config.minfds):
2508
+ close_fd(i)
2509
+
2510
+ def _spawn_as_child(self, filename: str, argv: ta.Sequence[str]) -> None:
2511
+ try:
2512
+ # prevent child from receiving signals sent to the parent by calling os.setpgrp to create a new process
2513
+ # group for the child; this prevents, for instance, the case of child processes being sent a SIGINT when
2514
+ # running supervisor in foreground mode and Ctrl-C in the terminal window running supervisord is pressed.
2515
+ # Presumably it also prevents HUP, etc received by supervisord from being sent to children.
2516
+ os.setpgrp()
2517
+
2518
+ self._prepare_child_fds()
2519
+ # sending to fd 2 will put this output in the stderr log
2520
+
2521
+ # set user
2522
+ setuid_msg = self.set_uid()
2523
+ if setuid_msg:
2524
+ uid = self.config.uid
2525
+ msg = f"couldn't setuid to {uid}: {setuid_msg}\n"
2526
+ os.write(2, as_bytes('supervisor: ' + msg))
2527
+ return # finally clause will exit the child process
2528
+
2529
+ # set environment
2530
+ env = os.environ.copy()
2531
+ env['SUPERVISOR_ENABLED'] = '1'
2532
+ env['SUPERVISOR_PROCESS_NAME'] = self.config.name
2533
+ if self.group:
2534
+ env['SUPERVISOR_GROUP_NAME'] = self.group.config.name
2535
+ if self.config.environment is not None:
2536
+ env.update(self.config.environment)
2537
+
2538
+ # change directory
2539
+ cwd = self.config.directory
2540
+ try:
2541
+ if cwd is not None:
2542
+ os.chdir(cwd)
2543
+ except OSError as why:
2544
+ code = errno.errorcode.get(why.args[0], why.args[0])
2545
+ msg = f"couldn't chdir to {cwd}: {code}\n"
2546
+ os.write(2, as_bytes('supervisor: ' + msg))
2547
+ return # finally clause will exit the child process
2548
+
2549
+ # set umask, then execve
2550
+ try:
2551
+ if self.config.umask is not None:
2552
+ os.umask(self.config.umask)
2553
+ os.execve(filename, list(argv), env)
2554
+ except OSError as why:
2555
+ code = errno.errorcode.get(why.args[0], why.args[0])
2556
+ msg = f"couldn't exec {argv[0]}: {code}\n"
2557
+ os.write(2, as_bytes('supervisor: ' + msg))
2558
+ except Exception: # noqa
2559
+ (file, fun, line), t, v, tbinfo = compact_traceback()
2560
+ error = f'{t}, {v}: file: {file} line: {line}'
2561
+ msg = f"couldn't exec {filename}: {error}\n"
2562
+ os.write(2, as_bytes('supervisor: ' + msg))
2563
+
2564
+ # this point should only be reached if execve failed. the finally clause will exit the child process.
2565
+
2566
+ finally:
2567
+ os.write(2, as_bytes('supervisor: child process was not spawned\n'))
2568
+ real_exit(127) # exit process with code for spawn failure
2569
+
2570
+ def _check_and_adjust_for_system_clock_rollback(self, test_time):
2571
+ """
2572
+ Check if system clock has rolled backward beyond test_time. If so, set affected timestamps to test_time.
2573
+ """
2574
+ if self.state == ProcessStates.STARTING:
2575
+ self.laststart = min(test_time, self.laststart)
2576
+ if self.delay > 0 and test_time < (self.delay - self.config.startsecs):
2577
+ self.delay = test_time + self.config.startsecs
2578
+
2579
+ elif self.state == ProcessStates.RUNNING:
2580
+ if test_time > self.laststart and test_time < (self.laststart + self.config.startsecs):
2581
+ self.laststart = test_time - self.config.startsecs
2582
+
2583
+ elif self.state == ProcessStates.STOPPING:
2584
+ self.last_stop_report = min(test_time, self.last_stop_report)
2585
+ if self.delay > 0 and test_time < (self.delay - self.config.stopwaitsecs):
2586
+ self.delay = test_time + self.config.stopwaitsecs
2587
+
2588
+ elif self.state == ProcessStates.BACKOFF:
2589
+ if self.delay > 0 and test_time < (self.delay - self.backoff):
2590
+ self.delay = test_time + self.backoff
2591
+
2592
+ def stop(self) -> ta.Optional[str]:
2593
+ self.administrative_stop = True
2594
+ self.last_stop_report = 0
2595
+ return self.kill(self.config.stopsignal)
2596
+
2597
+ def stop_report(self) -> None:
2598
+ """ Log a 'waiting for x to stop' message with throttling. """
2599
+ if self.state == ProcessStates.STOPPING:
2600
+ now = time.time()
2601
+
2602
+ self._check_and_adjust_for_system_clock_rollback(now)
2603
+
2604
+ if now > (self.last_stop_report + 2): # every 2 seconds
2605
+ log.info('waiting for %s to stop', as_string(self.config.name))
2606
+ self.last_stop_report = now
2607
+
2608
+ def give_up(self) -> None:
2609
+ self.delay = 0
2610
+ self.backoff = 0
2611
+ self.system_stop = True
2612
+ self._check_in_state(ProcessStates.BACKOFF)
2613
+ self.change_state(ProcessStates.FATAL)
2614
+
2615
+ def kill(self, sig: int) -> ta.Optional[str]:
2616
+ """
2617
+ Send a signal to the subprocess with the intention to kill it (to make it exit). This may or may not actually
2618
+ kill it.
2619
+
2620
+ Return None if the signal was sent, or an error message string if an error occurred or if the subprocess is not
2621
+ running.
2622
+ """
2623
+ now = time.time()
2624
+
2625
+ processname = as_string(self.config.name)
2626
+ # If the process is in BACKOFF and we want to stop or kill it, then BACKOFF -> STOPPED. This is needed because
2627
+ # if startretries is a large number and the process isn't starting successfully, the stop request would be
2628
+ # blocked for a long time waiting for the retries.
2629
+ if self.state == ProcessStates.BACKOFF:
2630
+ log.debug('Attempted to kill %s, which is in BACKOFF state.', processname)
2631
+ self.change_state(ProcessStates.STOPPED)
2632
+ return None
2633
+
2634
+ args: tuple
2635
+ if not self.pid:
2636
+ fmt, args = "attempted to kill %s with sig %s but it wasn't running", (processname, signame(sig))
2637
+ log.debug(fmt, *args)
2638
+ return fmt % args
2639
+
2640
+ # If we're in the stopping state, then we've already sent the stop signal and this is the kill signal
2641
+ if self.state == ProcessStates.STOPPING:
2642
+ killasgroup = self.config.killasgroup
2643
+ else:
2644
+ killasgroup = self.config.stopasgroup
2645
+
2646
+ as_group = ''
2647
+ if killasgroup:
2648
+ as_group = 'process group '
2649
+
2650
+ log.debug('killing %s (pid %s) %swith signal %s', processname, self.pid, as_group, signame(sig))
2651
+
2652
+ # RUNNING/STARTING/STOPPING -> STOPPING
2653
+ self.killing = True
2654
+ self.delay = now + self.config.stopwaitsecs
2655
+ # we will already be in the STOPPING state if we're doing a SIGKILL as a result of overrunning stopwaitsecs
2656
+ self._check_in_state(ProcessStates.RUNNING, ProcessStates.STARTING, ProcessStates.STOPPING)
2657
+ self.change_state(ProcessStates.STOPPING)
2658
+
2659
+ pid = self.pid
2660
+ if killasgroup:
2661
+ # send to the whole process group instead
2662
+ pid = -self.pid
2663
+
2664
+ try:
2665
+ try:
2666
+ os.kill(pid, sig)
2667
+ except OSError as exc:
2668
+ if exc.errno == errno.ESRCH:
2669
+ log.debug('unable to signal %s (pid %s), it probably just exited on its own: %s', processname, self.pid, str(exc)) # noqa
2670
+ # we could change the state here but we intentionally do not. we will do it during normal SIGCHLD
2671
+ # processing.
2672
+ return None
2673
+ raise
2674
+ except Exception: # noqa
2675
+ tb = traceback.format_exc()
2676
+ fmt, args = 'unknown problem killing %s (%s):%s', (processname, self.pid, tb)
2677
+ log.critical(fmt, *args)
2678
+ self.change_state(ProcessStates.UNKNOWN)
2679
+ self.killing = False
2680
+ self.delay = 0
2681
+ return fmt % args
2682
+
2683
+ return None
2684
+
2685
+ def signal(self, sig: int) -> ta.Optional[str]:
2686
+ """
2687
+ Send a signal to the subprocess, without intending to kill it.
2688
+
2689
+ Return None if the signal was sent, or an error message string if an error occurred or if the subprocess is not
2690
+ running.
2691
+ """
2692
+ processname = as_string(self.config.name)
2693
+ args: tuple
2694
+ if not self.pid:
2695
+ fmt, args = "attempted to send %s sig %s but it wasn't running", (processname, signame(sig))
2696
+ log.debug(fmt, *args)
2697
+ return fmt % args
2698
+
2699
+ log.debug('sending %s (pid %s) sig %s', processname, self.pid, signame(sig))
2700
+
2701
+ self._check_in_state(ProcessStates.RUNNING, ProcessStates.STARTING, ProcessStates.STOPPING)
2702
+
2703
+ try:
2704
+ try:
2705
+ os.kill(self.pid, sig)
2706
+ except OSError as exc:
2707
+ if exc.errno == errno.ESRCH:
2708
+ log.debug('unable to signal %s (pid %s), it probably just now exited '
2709
+ 'on its own: %s', processname, self.pid, str(exc))
2710
+ # we could change the state here but we intentionally do not. we will do it during normal SIGCHLD
2711
+ # processing.
2712
+ return None
2713
+ raise
2714
+ except Exception: # noqa
2715
+ tb = traceback.format_exc()
2716
+ fmt, args = 'unknown problem sending sig %s (%s):%s', (processname, self.pid, tb)
2717
+ log.critical(fmt, *args)
2718
+ self.change_state(ProcessStates.UNKNOWN)
2719
+ return fmt % args
2720
+
2721
+ return None
2722
+
2723
+ def finish(self, sts: int) -> None:
2724
+ """ The process was reaped and we need to report and manage its state """
2725
+ self.drain()
2726
+
2727
+ es, msg = decode_wait_status(sts)
2728
+
2729
+ now = time.time()
2730
+
2731
+ self._check_and_adjust_for_system_clock_rollback(now)
2732
+
2733
+ self.laststop = now
2734
+ processname = as_string(self.config.name)
2735
+
2736
+ if now > self.laststart:
2737
+ too_quickly = now - self.laststart < self.config.startsecs
2738
+ else:
2739
+ too_quickly = False
2740
+ log.warning(
2741
+ "process '%s' (%s) laststart time is in the future, don't "
2742
+ "know how long process was running so assuming it did "
2743
+ "not exit too quickly", processname, self.pid)
2744
+
2745
+ exit_expected = es in self.config.exitcodes
2746
+
2747
+ if self.killing:
2748
+ # likely the result of a stop request implies STOPPING -> STOPPED
2749
+ self.killing = False
2750
+ self.delay = 0
2751
+ self.exitstatus = es
2752
+
2753
+ fmt, args = 'stopped: %s (%s)', (processname, msg)
2754
+ self._check_in_state(ProcessStates.STOPPING)
2755
+ self.change_state(ProcessStates.STOPPED)
2756
+ if exit_expected:
2757
+ log.info(fmt, *args)
2758
+ else:
2759
+ log.warning(fmt, *args)
2760
+
2761
+ elif too_quickly:
2762
+ # the program did not stay up long enough to make it to RUNNING implies STARTING -> BACKOFF
2763
+ self.exitstatus = None
2764
+ self.spawn_err = 'Exited too quickly (process log may have details)'
2765
+ self._check_in_state(ProcessStates.STARTING)
2766
+ self.change_state(ProcessStates.BACKOFF)
2767
+ log.warning('exited: %s (%s)', processname, msg + '; not expected')
2768
+
2769
+ else:
2770
+ # this finish was not the result of a stop request, the program was in the RUNNING state but exited implies
2771
+ # RUNNING -> EXITED normally but see next comment
2772
+ self.delay = 0
2773
+ self.backoff = 0
2774
+ self.exitstatus = es
2775
+
2776
+ # if the process was STARTING but a system time change causes self.laststart to be in the future, the normal
2777
+ # STARTING->RUNNING transition can be subverted so we perform the transition here.
2778
+ if self.state == ProcessStates.STARTING:
2779
+ self.change_state(ProcessStates.RUNNING)
2780
+
2781
+ self._check_in_state(ProcessStates.RUNNING)
2782
+
2783
+ if exit_expected:
2784
+ # expected exit code
2785
+ self.change_state(ProcessStates.EXITED, expected=True)
2786
+ log.info('exited: %s (%s)', processname, msg + '; expected')
2787
+ else:
2788
+ # unexpected exit code
2789
+ self.spawn_err = f'Bad exit code {es}'
2790
+ self.change_state(ProcessStates.EXITED, expected=False)
2791
+ log.warning('exited: %s (%s)', processname, msg + '; not expected')
2792
+
2793
+ self._pid = 0
2794
+ close_parent_pipes(self._pipes)
2795
+ self._pipes = {}
2796
+ self._dispatchers = {}
2797
+
2798
+ # if we died before we processed the current event (only happens if we're an event listener), notify the event
2799
+ # system that this event was rejected so it can be processed again.
2800
+ if self.event is not None:
2801
+ # Note: this should only be true if we were in the BUSY state when finish() was called.
2802
+ notify_event(EventRejectedEvent(self, self.event)) # type: ignore
2803
+ self.event = None
2804
+
2805
+ def set_uid(self) -> ta.Optional[str]:
2806
+ if self.config.uid is None:
2807
+ return None
2808
+ msg = drop_privileges(self.config.uid)
2809
+ return msg
2810
+
2811
+ def __lt__(self, other):
2812
+ return self.config.priority < other.config.priority
2813
+
2814
+ def __eq__(self, other):
2815
+ return self.config.priority == other.config.priority
2816
+
2817
+ def __repr__(self):
2818
+ # repr can't return anything other than a native string, but the name might be unicode - a problem on Python 2.
2819
+ name = self.config.name
2820
+ return f'<Subprocess at {id(self)} with name {name} in state {get_process_state_description(self.get_state())}>'
2821
+
2822
+ def get_state(self) -> ProcessState:
2823
+ return self.state
2824
+
2825
+ def transition(self):
2826
+ now = time.time()
2827
+ state = self.state
2828
+
2829
+ self._check_and_adjust_for_system_clock_rollback(now)
2830
+
2831
+ logger = log
2832
+
2833
+ if self.context.state > SupervisorStates.RESTARTING:
2834
+ # dont start any processes if supervisor is shutting down
2835
+ if state == ProcessStates.EXITED:
2836
+ if self.config.autorestart:
2837
+ if self.config.autorestart is RestartUnconditionally:
2838
+ # EXITED -> STARTING
2839
+ self.spawn()
2840
+ elif self.exitstatus not in self.config.exitcodes: # type: ignore
2841
+ # EXITED -> STARTING
2842
+ self.spawn()
2843
+
2844
+ elif state == ProcessStates.STOPPED and not self.laststart:
2845
+ if self.config.autostart:
2846
+ # STOPPED -> STARTING
2847
+ self.spawn()
2848
+
2849
+ elif state == ProcessStates.BACKOFF:
2850
+ if self.backoff <= self.config.startretries:
2851
+ if now > self.delay:
2852
+ # BACKOFF -> STARTING
2853
+ self.spawn()
2854
+
2855
+ processname = as_string(self.config.name)
2856
+ if state == ProcessStates.STARTING:
2857
+ if now - self.laststart > self.config.startsecs:
2858
+ # STARTING -> RUNNING if the proc has started successfully and it has stayed up for at least
2859
+ # proc.config.startsecs,
2860
+ self.delay = 0
2861
+ self.backoff = 0
2862
+ self._check_in_state(ProcessStates.STARTING)
2863
+ self.change_state(ProcessStates.RUNNING)
2864
+ msg = ('entered RUNNING state, process has stayed up for > than %s seconds (startsecs)' % self.config.startsecs) # noqa
2865
+ logger.info('success: %s %s', processname, msg)
2866
+
2867
+ if state == ProcessStates.BACKOFF:
2868
+ if self.backoff > self.config.startretries:
2869
+ # BACKOFF -> FATAL if the proc has exceeded its number of retries
2870
+ self.give_up()
2871
+ msg = ('entered FATAL state, too many start retries too quickly')
2872
+ logger.info('gave up: %s %s', processname, msg)
2873
+
2874
+ elif state == ProcessStates.STOPPING:
2875
+ time_left = self.delay - now
2876
+ if time_left <= 0:
2877
+ # kill processes which are taking too long to stop with a final sigkill. if this doesn't kill it, the
2878
+ # process will be stuck in the STOPPING state forever.
2879
+ log.warning('killing \'%s\' (%s) with SIGKILL', processname, self.pid)
2880
+ self.kill(signal.SIGKILL)
2881
+
2882
+ def create_auto_child_logs(self):
2883
+ # temporary logfiles which are erased at start time
2884
+ # get_autoname = self.context.get_auto_child_log_name # noqa
2885
+ # sid = self.context.config.identifier # noqa
2886
+ # name = self.config.name # noqa
2887
+ # if self.stdout_logfile is Automatic:
2888
+ # self.stdout_logfile = get_autoname(name, sid, 'stdout')
2889
+ # if self.stderr_logfile is Automatic:
2890
+ # self.stderr_logfile = get_autoname(name, sid, 'stderr')
2891
+ pass
2892
+
2893
+
2894
+ @functools.total_ordering
2895
+ class ProcessGroup:
2896
+ def __init__(self, config: ProcessGroupConfig, context: ServerContext):
2897
+ super().__init__()
2898
+ self.config = config
2899
+ self.context = context
2900
+ self.processes = {}
2901
+ for pconfig in self.config.processes or []:
2902
+ process = Subprocess(pconfig, self, self.context)
2903
+ self.processes[pconfig.name] = process
2904
+
2905
+ def __lt__(self, other):
2906
+ return self.config.priority < other.config.priority
2907
+
2908
+ def __eq__(self, other):
2909
+ return self.config.priority == other.config.priority
2910
+
2911
+ def __repr__(self):
2912
+ # repr can't return anything other than a native string, but the name might be unicode - a problem on Python 2.
2913
+ name = self.config.name
2914
+ return f'<{self.__class__.__name__} instance at {id(self)} named {name}>'
2915
+
2916
+ def remove_logs(self) -> None:
2917
+ for process in self.processes.values():
2918
+ process.remove_logs()
2919
+
2920
+ def reopen_logs(self) -> None:
2921
+ for process in self.processes.values():
2922
+ process.reopen_logs()
2923
+
2924
+ def stop_all(self) -> None:
2925
+ processes = list(self.processes.values())
2926
+ processes.sort()
2927
+ processes.reverse() # stop in desc priority order
2928
+
2929
+ for proc in processes:
2930
+ state = proc.get_state()
2931
+ if state == ProcessStates.RUNNING:
2932
+ # RUNNING -> STOPPING
2933
+ proc.stop()
2934
+
2935
+ elif state == ProcessStates.STARTING:
2936
+ # STARTING -> STOPPING
2937
+ proc.stop()
2938
+
2939
+ elif state == ProcessStates.BACKOFF:
2940
+ # BACKOFF -> FATAL
2941
+ proc.give_up()
2942
+
2943
+ def get_unstopped_processes(self) -> ta.List[Subprocess]:
2944
+ return [x for x in self.processes.values() if x.get_state() not in STOPPED_STATES]
2945
+
2946
+ def get_dispatchers(self) -> ta.Dict[int, Dispatcher]:
2947
+ dispatchers = {}
2948
+ for process in self.processes.values():
2949
+ dispatchers.update(process._dispatchers) # noqa
2950
+ return dispatchers
2951
+
2952
+ def before_remove(self) -> None:
2953
+ pass
2954
+
2955
+ def transition(self) -> None:
2956
+ for proc in self.processes.values():
2957
+ proc.transition()
2958
+
2959
+ def after_setuid(self) -> None:
2960
+ for proc in self.processes.values():
2961
+ proc.create_auto_child_logs()
2962
+
2963
+
2964
+ ########################################
2965
+ # supervisor.py
2966
+
2967
+
2968
+ log = logging.getLogger(__name__)
2969
+
2970
+
2971
+ class Supervisor:
2972
+
2973
+ def __init__(self, context: ServerContext) -> None:
2974
+ super().__init__()
2975
+
2976
+ self._context = context
2977
+ self._ticks: ta.Dict[int, float] = {}
2978
+ self._process_groups: ta.Dict[str, ProcessGroup] = {} # map of process group name to process group object
2979
+ self._stop_groups: ta.Optional[ta.List[ProcessGroup]] = None # list used for priority ordered shutdown
2980
+ self._stopping = False # set after we detect that we are handling a stop request
2981
+ self._last_shutdown_report = 0. # throttle for delayed process error reports at stop
2982
+
2983
+ @property
2984
+ def context(self) -> ServerContext:
2985
+ return self._context
2986
+
2987
+ def get_state(self) -> SupervisorState:
2988
+ return self._context.state
2989
+
2990
+ def main(self) -> None:
2991
+ if not self._context.first:
2992
+ # prevent crash on libdispatch-based systems, at least for the first request
2993
+ self._context.cleanup_fds()
2994
+
2995
+ self._context.set_uid_or_exit()
2996
+
2997
+ if self._context.first:
2998
+ self._context.set_rlimits_or_exit()
2999
+
3000
+ # this sets the options.logger object delay logger instantiation until after setuid
3001
+ if not self._context.config.nocleanup:
3002
+ # clean up old automatic logs
3003
+ self._context.clear_auto_child_logdir()
3004
+
3005
+ self.run()
3006
+
3007
+ def run(self) -> None:
3008
+ self._process_groups = {} # clear
3009
+ self._stop_groups = None # clear
3010
+
3011
+ clear_events()
3012
+
3013
+ try:
3014
+ for config in self._context.config.groups or []:
3015
+ self.add_process_group(config)
3016
+
3017
+ self._context.set_signals()
3018
+
3019
+ if not self._context.config.nodaemon and self._context.first:
3020
+ self._context.daemonize()
3021
+
3022
+ # writing pid file needs to come *after* daemonizing or pid will be wrong
3023
+ self._context.write_pidfile()
3024
+
3025
+ self.runforever()
3026
+
3027
+ finally:
3028
+ self._context.cleanup()
3029
+
3030
+ def diff_to_active(self):
3031
+ new = self._context.config.groups or []
3032
+ cur = [group.config for group in self._process_groups.values()]
3033
+
3034
+ curdict = dict(zip([cfg.name for cfg in cur], cur))
3035
+ newdict = dict(zip([cfg.name for cfg in new], new))
3036
+
3037
+ added = [cand for cand in new if cand.name not in curdict]
3038
+ removed = [cand for cand in cur if cand.name not in newdict]
3039
+
3040
+ changed = [cand for cand in new if cand != curdict.get(cand.name, cand)]
3041
+
3042
+ return added, changed, removed
3043
+
3044
+ def add_process_group(self, config: ProcessGroupConfig) -> bool:
3045
+ name = config.name
3046
+ if name in self._process_groups:
3047
+ return False
3048
+
3049
+ group = self._process_groups[name] = ProcessGroup(config, self._context)
3050
+ group.after_setuid()
3051
+
3052
+ notify_event(ProcessGroupAddedEvent(name))
3053
+ return True
3054
+
3055
+ def remove_process_group(self, name: str) -> bool:
3056
+ if self._process_groups[name].get_unstopped_processes():
3057
+ return False
3058
+
3059
+ self._process_groups[name].before_remove()
3060
+
3061
+ del self._process_groups[name]
3062
+
3063
+ notify_event(ProcessGroupRemovedEvent(name))
3064
+ return True
3065
+
3066
+ def get_process_map(self) -> ta.Dict[int, Dispatcher]:
3067
+ process_map = {}
3068
+ for group in self._process_groups.values():
3069
+ process_map.update(group.get_dispatchers())
3070
+ return process_map
3071
+
3072
+ def shutdown_report(self) -> ta.List[Subprocess]:
3073
+ unstopped: ta.List[Subprocess] = []
3074
+
3075
+ for group in self._process_groups.values():
3076
+ unstopped.extend(group.get_unstopped_processes())
3077
+
3078
+ if unstopped:
3079
+ # throttle 'waiting for x to die' reports
3080
+ now = time.time()
3081
+ if now > (self._last_shutdown_report + 3): # every 3 secs
3082
+ names = [as_string(p.config.name) for p in unstopped]
3083
+ namestr = ', '.join(names)
3084
+ log.info('waiting for %s to die', namestr)
3085
+ self._last_shutdown_report = now
3086
+ for proc in unstopped:
3087
+ state = get_process_state_description(proc.get_state())
3088
+ log.debug('%s state: %s', proc.config.name, state)
3089
+
3090
+ return unstopped
3091
+
3092
+ def _ordered_stop_groups_phase_1(self) -> None:
3093
+ if self._stop_groups:
3094
+ # stop the last group (the one with the "highest" priority)
3095
+ self._stop_groups[-1].stop_all()
3096
+
3097
+ def _ordered_stop_groups_phase_2(self) -> None:
3098
+ # after phase 1 we've transitioned and reaped, let's see if we can remove the group we stopped from the
3099
+ # stop_groups queue.
3100
+ if self._stop_groups:
3101
+ # pop the last group (the one with the "highest" priority)
3102
+ group = self._stop_groups.pop()
3103
+ if group.get_unstopped_processes():
3104
+ # if any processes in the group aren't yet in a stopped state, we're not yet done shutting this group
3105
+ # down, so push it back on to the end of the stop group queue
3106
+ self._stop_groups.append(group)
3107
+
3108
+ def runforever(self) -> None:
3109
+ notify_event(SupervisorRunningEvent())
3110
+ timeout = 1 # this cannot be fewer than the smallest TickEvent (5)
3111
+
3112
+ while True:
3113
+ combined_map = {}
3114
+ combined_map.update(self.get_process_map())
3115
+
3116
+ pgroups = list(self._process_groups.values())
3117
+ pgroups.sort()
3118
+
3119
+ if self._context.state < SupervisorStates.RUNNING:
3120
+ if not self._stopping:
3121
+ # first time, set the stopping flag, do a notification and set stop_groups
3122
+ self._stopping = True
3123
+ self._stop_groups = pgroups[:]
3124
+ notify_event(SupervisorStoppingEvent())
3125
+
3126
+ self._ordered_stop_groups_phase_1()
3127
+
3128
+ if not self.shutdown_report():
3129
+ # if there are no unstopped processes (we're done killing everything), it's OK to shutdown or reload
3130
+ raise ExitNow
3131
+
3132
+ for fd, dispatcher in combined_map.items():
3133
+ if dispatcher.readable():
3134
+ self._context.poller.register_readable(fd)
3135
+ if dispatcher.writable():
3136
+ self._context.poller.register_writable(fd)
3137
+
3138
+ r, w = self._context.poller.poll(timeout)
3139
+
3140
+ for fd in r:
3141
+ if fd in combined_map:
3142
+ try:
3143
+ dispatcher = combined_map[fd]
3144
+ log.debug('read event caused by %r', dispatcher)
3145
+ dispatcher.handle_read_event()
3146
+ if not dispatcher.readable():
3147
+ self._context.poller.unregister_readable(fd)
3148
+ except ExitNow:
3149
+ raise
3150
+ except Exception: # noqa
3151
+ combined_map[fd].handle_error()
3152
+ else:
3153
+ # if the fd is not in combined_map, we should unregister it. otherwise, it will be polled every
3154
+ # time, which may cause 100% cpu usage
3155
+ log.debug('unexpected read event from fd %r', fd)
3156
+ try:
3157
+ self._context.poller.unregister_readable(fd)
3158
+ except Exception: # noqa
3159
+ pass
3160
+
3161
+ for fd in w:
3162
+ if fd in combined_map:
3163
+ try:
3164
+ dispatcher = combined_map[fd]
3165
+ log.debug('write event caused by %r', dispatcher)
3166
+ dispatcher.handle_write_event()
3167
+ if not dispatcher.writable():
3168
+ self._context.poller.unregister_writable(fd)
3169
+ except ExitNow:
3170
+ raise
3171
+ except Exception: # noqa
3172
+ combined_map[fd].handle_error()
3173
+ else:
3174
+ log.debug('unexpected write event from fd %r', fd)
3175
+ try:
3176
+ self._context.poller.unregister_writable(fd)
3177
+ except Exception: # noqa
3178
+ pass
3179
+
3180
+ for group in pgroups:
3181
+ group.transition()
3182
+
3183
+ self._reap()
3184
+ self._handle_signal()
3185
+ self._tick()
3186
+
3187
+ if self._context.state < SupervisorStates.RUNNING:
3188
+ self._ordered_stop_groups_phase_2()
3189
+
3190
+ if self._context.test:
3191
+ break
3192
+
3193
+ def _tick(self, now: ta.Optional[float] = None) -> None:
3194
+ """Send one or more 'tick' events when the timeslice related to the period for the event type rolls over"""
3195
+
3196
+ if now is None:
3197
+ # now won't be None in unit tests
3198
+ now = time.time()
3199
+
3200
+ for event in TICK_EVENTS:
3201
+ period = event.period # type: ignore
3202
+
3203
+ last_tick = self._ticks.get(period)
3204
+ if last_tick is None:
3205
+ # we just started up
3206
+ last_tick = self._ticks[period] = timeslice(period, now)
3207
+
3208
+ this_tick = timeslice(period, now)
3209
+ if this_tick != last_tick:
3210
+ self._ticks[period] = this_tick
3211
+ notify_event(event(this_tick, self))
3212
+
3213
+ def _reap(self, *, once: bool = False, depth: int = 0) -> None:
3214
+ if depth >= 100:
3215
+ return
3216
+
3217
+ pid, sts = self._context.waitpid()
3218
+ if not pid:
3219
+ return
3220
+
3221
+ process = self._context.pid_history.get(pid, None)
3222
+ if process is None:
3223
+ _, msg = decode_wait_status(check_not_none(sts))
3224
+ log.info('reaped unknown pid %s (%s)', pid, msg)
3225
+ else:
3226
+ process.finish(check_not_none(sts))
3227
+ del self._context.pid_history[pid]
3228
+
3229
+ if not once:
3230
+ # keep reaping until no more kids to reap, but don't recurse infinitely
3231
+ self._reap(once=False, depth=depth + 1)
3232
+
3233
+ def _handle_signal(self) -> None:
3234
+ sig = self._context.get_signal()
3235
+ if not sig:
3236
+ return
3237
+
3238
+ if sig in (signal.SIGTERM, signal.SIGINT, signal.SIGQUIT):
3239
+ log.warning('received %s indicating exit request', signame(sig))
3240
+ self._context.set_state(SupervisorStates.SHUTDOWN)
3241
+
3242
+ elif sig == signal.SIGHUP:
3243
+ if self._context.state == SupervisorStates.SHUTDOWN:
3244
+ log.warning('ignored %s indicating restart request (shutdown in progress)', signame(sig)) # noqa
3245
+ else:
3246
+ log.warning('received %s indicating restart request', signame(sig)) # noqa
3247
+ self._context.set_state(SupervisorStates.RESTARTING)
3248
+
3249
+ elif sig == signal.SIGCHLD:
3250
+ log.debug('received %s indicating a child quit', signame(sig))
3251
+
3252
+ elif sig == signal.SIGUSR2:
3253
+ log.info('received %s indicating log reopen request', signame(sig))
3254
+ # self._context.reopen_logs()
3255
+ for group in self._process_groups.values():
3256
+ group.reopen_logs()
3257
+
3258
+ else:
3259
+ log.debug('received %s indicating nothing', signame(sig))
3260
+
3261
+
3262
+ def timeslice(period, when):
3263
+ return int(when - (when % period))
3264
+
3265
+
3266
+ def main(args=None, test=False):
3267
+ configure_standard_logging('INFO')
3268
+
3269
+ # if we hup, restart by making a new Supervisor()
3270
+ first = True
3271
+ while True:
3272
+ config = ServerConfig.new(
3273
+ nodaemon=True,
3274
+ groups=[
3275
+ ProcessGroupConfig(
3276
+ name='default',
3277
+ processes=[
3278
+ ProcessConfig(
3279
+ name='sleep',
3280
+ command='sleep 600',
3281
+ stdout=ProcessConfig.Log(
3282
+ file='/dev/fd/1',
3283
+ maxbytes=0,
3284
+ ),
3285
+ redirect_stderr=True,
3286
+ ),
3287
+ ProcessConfig(
3288
+ name='ls',
3289
+ command='ls -al',
3290
+ stdout=ProcessConfig.Log(
3291
+ file='/dev/fd/1',
3292
+ maxbytes=0,
3293
+ ),
3294
+ redirect_stderr=True,
3295
+ ),
3296
+ ],
3297
+ ),
3298
+ ],
3299
+ )
3300
+
3301
+ context = ServerContext(
3302
+ config,
3303
+ )
3304
+
3305
+ context.first = first
3306
+ context.test = test
3307
+ go(context)
3308
+ # options.close_logger()
3309
+ first = False
3310
+ if test or (context.state < SupervisorStates.RESTARTING):
3311
+ break
3312
+
3313
+
3314
+ def go(context): # pragma: no cover
3315
+ d = Supervisor(context)
3316
+ try:
3317
+ d.main()
3318
+ except ExitNow:
3319
+ pass
3320
+
3321
+
3322
+ if __name__ == '__main__':
3323
+ main()