projects
/
python_utils.git
/ blobdiff
commit
grep
author
committer
pickaxe
?
search:
re
summary
|
shortlog
|
log
|
commit
|
commitdiff
|
tree
raw
|
inline
| side by side
Add a simple test to google_assistant code.
[python_utils.git]
/
logging_utils.py
diff --git
a/logging_utils.py
b/logging_utils.py
index 3fd2b849a3dcc2d030de28e0abe6c70cb1d9fe02..ca1544150065b367b976d9ca43382fa1a6cff5e5 100644
(file)
--- a/
logging_utils.py
+++ b/
logging_utils.py
@@
-1,4
+1,5
@@
#!/usr/bin/env python3
#!/usr/bin/env python3
+# -*- coding: utf-8 -*-
"""Utilities related to logging."""
"""Utilities related to logging."""
@@
-223,8
+224,8
@@
class SquelchRepeatedMessagesFilter(logging.Filter):
"""
def __init__(self) -> None:
"""
def __init__(self) -> None:
- self.counters: collections.Counter = collections.Counter()
super().__init__()
super().__init__()
+ self.counters: collections.Counter = collections.Counter()
@overrides
def filter(self, record: logging.LogRecord) -> bool:
@overrides
def filter(self, record: logging.LogRecord) -> bool:
@@
-257,9
+258,7
@@
class DynamicPerScopeLoggingLevelFilter(logging.Filter):
per_scope_logging_levels: str,
) -> None:
super().__init__()
per_scope_logging_levels: str,
) -> None:
super().__init__()
- self.valid_levels = set(
- ['NOTSET', 'DEBUG', 'INFO', 'WARNING', 'ERROR', 'CRITICAL']
- )
+ self.valid_levels = set(['NOTSET', 'DEBUG', 'INFO', 'WARNING', 'ERROR', 'CRITICAL'])
self.default_logging_level = default_logging_level
self.level_by_scope = {}
if per_scope_logging_levels is not None:
self.default_logging_level = default_logging_level
self.level_by_scope = {}
if per_scope_logging_levels is not None:
@@
-286,9
+285,9
@@
class DynamicPerScopeLoggingLevelFilter(logging.Filter):
file=sys.stderr,
)
continue
file=sys.stderr,
)
continue
- self.level_by_scope[
- scope
-
] = DynamicPerScopeLoggingLevelFilter.level_name_to_level(level
)
+ self.level_by_scope[
scope] = DynamicPerScopeLoggingLevelFilter.level_name_to_level(
+ level
+ )
@overrides
def filter(self, record: logging.LogRecord) -> bool:
@overrides
def filter(self, record: logging.LogRecord) -> bool:
@@
-386,9
+385,7
@@
class MillisecondAwareFormatter(logging.Formatter):
@overrides
def formatTime(self, record, datefmt=None):
@overrides
def formatTime(self, record, datefmt=None):
- ct = MillisecondAwareFormatter.converter(
- record.created, pytz.timezone("US/Pacific")
- )
+ ct = MillisecondAwareFormatter.converter(record.created, pytz.timezone("US/Pacific"))
if datefmt:
s = ct.strftime(datefmt)
else:
if datefmt:
s = ct.strftime(datefmt)
else:
@@
-398,36
+395,29
@@
class MillisecondAwareFormatter(logging.Formatter):
def log_about_logging(
def log_about_logging(
- logger, default_logging_level, preexisting_handlers_count, fmt, facility_name
+ logger,
+ default_logging_level,
+ preexisting_handlers_count,
+ fmt,
+ facility_name,
):
):
- level_name = logging._levelToName.get(
- default_logging_level, str(default_logging_level)
- )
+ level_name = logging._levelToName.get(default_logging_level, str(default_logging_level))
logger.debug(f'Initialized global logging; default logging level is {level_name}.')
logger.debug(f'Initialized global logging; default logging level is {level_name}.')
- if (
- config.config['logging_clear_preexisting_handlers']
- and preexisting_handlers_count > 0
- ):
+ if config.config['logging_clear_preexisting_handlers'] and preexisting_handlers_count > 0:
msg = f'Logging cleared {preexisting_handlers_count} global handlers (--logging_clear_preexisting_handlers)'
logger.warning(msg)
logger.debug(f'Logging format specification is "{fmt}"')
if config.config['logging_debug_threads']:
msg = f'Logging cleared {preexisting_handlers_count} global handlers (--logging_clear_preexisting_handlers)'
logger.warning(msg)
logger.debug(f'Logging format specification is "{fmt}"')
if config.config['logging_debug_threads']:
- logger.debug(
- '...Logging format spec captures tid/pid (--logging_debug_threads)'
- )
+ logger.debug('...Logging format spec captures tid/pid (--logging_debug_threads)')
if config.config['logging_debug_modules']:
logger.debug(
'...Logging format spec captures files/functions/lineno (--logging_debug_modules)'
)
if config.config['logging_syslog']:
if config.config['logging_debug_modules']:
logger.debug(
'...Logging format spec captures files/functions/lineno (--logging_debug_modules)'
)
if config.config['logging_syslog']:
- logger.debug(
- f'Logging to syslog as {facility_name} with priority mapping based on level'
- )
+ logger.debug(f'Logging to syslog as {facility_name} with priority mapping based on level')
if config.config['logging_filename']:
logger.debug(f'Logging to filename {config.config["logging_filename"]}')
if config.config['logging_filename']:
logger.debug(f'Logging to filename {config.config["logging_filename"]}')
- logger.debug(
- f'...with {config.config["logging_filename_maxsize"]} bytes max file size.'
- )
+ logger.debug(f'...with {config.config["logging_filename_maxsize"]} bytes max file size.')
logger.debug(
f'...and {config.config["logging_filename_count"]} rotating backup file count.'
)
logger.debug(
f'...and {config.config["logging_filename_count"]} rotating backup file count.'
)
@@
-487,9
+477,7
@@
def initialize_logging(logger=None) -> logging.Logger:
handler: Optional[logging.Handler] = None
# Global default logging level (--logging_level)
handler: Optional[logging.Handler] = None
# Global default logging level (--logging_level)
- default_logging_level = getattr(
- logging, config.config['logging_level'].upper(), None
- )
+ default_logging_level = getattr(logging, config.config['logging_level'].upper(), None)
if not isinstance(default_logging_level, int):
raise ValueError('Invalid level: %s' % config.config['logging_level'])
if not isinstance(default_logging_level, int):
raise ValueError('Invalid level: %s' % config.config['logging_level'])