projects
/
python_utils.git
/ blobdiff
commit
grep
author
committer
pickaxe
?
search:
re
summary
|
shortlog
|
log
|
commit
|
commitdiff
|
tree
raw
|
inline
| side by side
More cleanup, yey!
[python_utils.git]
/
string_utils.py
diff --git
a/string_utils.py
b/string_utils.py
index 1ed9b4ad10790685290838cd30a2322d59b1b373..d75c6ba1aca2c559ed4254d535747c54f4719bf5 100644
(file)
--- a/
string_utils.py
+++ b/
string_utils.py
@@
-1,4
+1,5
@@
#!/usr/bin/env python3
#!/usr/bin/env python3
+# -*- coding: utf-8 -*-
"""The MIT License (MIT)
"""The MIT License (MIT)
@@
-72,9
+73,7
@@
URLS_RE = re.compile(r"({})".format(URLS_RAW_STRING), re.IGNORECASE)
ESCAPED_AT_SIGN = re.compile(r'(?!"[^"]*)@+(?=[^"]*")|\\@')
ESCAPED_AT_SIGN = re.compile(r'(?!"[^"]*)@+(?=[^"]*")|\\@')
-EMAILS_RAW_STRING = (
- r"[a-zA-Z\d._\+\-'`!%#$&*/=\?\^\{\}\|~\\]+@[a-z\d-]+\.?[a-z\d-]+\.[a-z]{2,4}"
-)
+EMAILS_RAW_STRING = r"[a-zA-Z\d._\+\-'`!%#$&*/=\?\^\{\}\|~\\]+@[a-z\d-]+\.?[a-z\d-]+\.[a-z]{2,4}"
EMAIL_RE = re.compile(r"^{}$".format(EMAILS_RAW_STRING))
EMAIL_RE = re.compile(r"^{}$".format(EMAILS_RAW_STRING))
@@
-84,13
+83,9
@@
CAMEL_CASE_TEST_RE = re.compile(r"^[a-zA-Z]*([a-z]+[A-Z]+|[A-Z]+[a-z]+)[a-zA-Z\d
CAMEL_CASE_REPLACE_RE = re.compile(r"([a-z]|[A-Z]+)(?=[A-Z])")
CAMEL_CASE_REPLACE_RE = re.compile(r"([a-z]|[A-Z]+)(?=[A-Z])")
-SNAKE_CASE_TEST_RE = re.compile(
- r"^([a-z]+\d*_[a-z\d_]*|_+[a-z\d]+[a-z\d_]*)$", re.IGNORECASE
-)
+SNAKE_CASE_TEST_RE = re.compile(r"^([a-z]+\d*_[a-z\d_]*|_+[a-z\d]+[a-z\d_]*)$", re.IGNORECASE)
-SNAKE_CASE_TEST_DASH_RE = re.compile(
- r"([a-z]+\d*-[a-z\d-]*|-+[a-z\d]+[a-z\d-]*)$", re.IGNORECASE
-)
+SNAKE_CASE_TEST_DASH_RE = re.compile(r"([a-z]+\d*-[a-z\d-]*|-+[a-z\d]+[a-z\d-]*)$", re.IGNORECASE)
SNAKE_CASE_REPLACE_RE = re.compile(r"(_)([a-z\d])")
SNAKE_CASE_REPLACE_RE = re.compile(r"(_)([a-z\d])")
@@
-107,9
+102,7
@@
CREDIT_CARDS = {
JSON_WRAPPER_RE = re.compile(r"^\s*[\[{]\s*(.*)\s*[\}\]]\s*$", re.MULTILINE | re.DOTALL)
JSON_WRAPPER_RE = re.compile(r"^\s*[\[{]\s*(.*)\s*[\}\]]\s*$", re.MULTILINE | re.DOTALL)
-UUID_RE = re.compile(
- r"^[a-f\d]{8}-[a-f\d]{4}-[a-f\d]{4}-[a-f\d]{4}-[a-f\d]{12}$", re.IGNORECASE
-)
+UUID_RE = re.compile(r"^[a-f\d]{8}-[a-f\d]{4}-[a-f\d]{4}-[a-f\d]{4}-[a-f\d]{12}$", re.IGNORECASE)
UUID_HEX_OK_RE = re.compile(
r"^[a-f\d]{8}-?[a-f\d]{4}-?[a-f\d]{4}-?[a-f\d]{4}-?[a-f\d]{12}$",
UUID_HEX_OK_RE = re.compile(
r"^[a-f\d]{8}-?[a-f\d]{4}-?[a-f\d]{4}-?[a-f\d]{4}-?[a-f\d]{12}$",
@@
-126,9
+119,7
@@
ANYWHERE_IP_V6_RE = re.compile(r"([a-z\d]{0,4}:){7}[a-z\d]{0,4}", re.IGNORECASE)
MAC_ADDRESS_RE = re.compile(r"^([0-9A-F]{2}[:-]){5}([0-9A-F]{2})$", re.IGNORECASE)
MAC_ADDRESS_RE = re.compile(r"^([0-9A-F]{2}[:-]){5}([0-9A-F]{2})$", re.IGNORECASE)
-ANYWHERE_MAC_ADDRESS_RE = re.compile(
- r"([0-9A-F]{2}[:-]){5}([0-9A-F]{2})", re.IGNORECASE
-)
+ANYWHERE_MAC_ADDRESS_RE = re.compile(r"([0-9A-F]{2}[:-]){5}([0-9A-F]{2})", re.IGNORECASE)
WORDS_COUNT_RE = re.compile(r"\W*[^\W_]+\W*", re.IGNORECASE | re.MULTILINE | re.UNICODE)
WORDS_COUNT_RE = re.compile(r"\W*[^\W_]+\W*", re.IGNORECASE | re.MULTILINE | re.UNICODE)
@@
-151,16
+142,16
@@
MARGIN_RE = re.compile(r"^[^\S\r\n]+")
ESCAPE_SEQUENCE_RE = re.compile(r"
\e
\[[^A-Za-z]*[A-Za-z]")
NUM_SUFFIXES = {
ESCAPE_SEQUENCE_RE = re.compile(r"
\e
\[[^A-Za-z]*[A-Za-z]")
NUM_SUFFIXES = {
- "Pb": (1024
**
5),
- "P": (1024
**
5),
- "Tb": (1024
**
4),
- "T": (1024
**
4),
- "Gb": (1024
**
3),
- "G": (1024
**
3),
- "Mb": (1024
**
2),
- "M": (1024
**
2),
- "Kb": (1024
**
1),
- "K": (1024
**
1),
+ "Pb": (1024
**
5),
+ "P": (1024
**
5),
+ "Tb": (1024
**
4),
+ "T": (1024
**
4),
+ "Gb": (1024
**
3),
+ "G": (1024
**
3),
+ "Mb": (1024
**
2),
+ "M": (1024
**
2),
+ "Kb": (1024
**
1),
+ "K": (1024
**
1),
}
}
@@
-418,9
+409,7
@@
def add_thousands_separator(in_str: str, *, separator_char=',', places=3) -> str
if isinstance(in_str, numbers.Number):
in_str = f'{in_str}'
if is_number(in_str):
if isinstance(in_str, numbers.Number):
in_str = f'{in_str}'
if is_number(in_str):
- return _add_thousands_separator(
- in_str, separator_char=separator_char, places=places
- )
+ return _add_thousands_separator(in_str, separator_char=separator_char, places=places)
raise ValueError(in_str)
raise ValueError(in_str)
@@
-1086,13
+1075,13
@@
def to_date(in_str: str) -> Optional[datetime.date]:
"""
Parses a date string. See DateParser docs for details.
"""
"""
Parses a date string. See DateParser docs for details.
"""
- import dateparse.dateparse_utils as d
p # type: ignore
+ import dateparse.dateparse_utils as d
u
try:
try:
- d = d
p.DateParser()
+ d = d
u.DateParser() # type: ignore
d.parse(in_str)
return d.get_date()
d.parse(in_str)
return d.get_date()
- except d
p.ParseException:
+ except d
u.ParseException: # type: ignore
msg = f'Unable to parse date {in_str}.'
logger.warning(msg)
return None
msg = f'Unable to parse date {in_str}.'
logger.warning(msg)
return None
@@
-1105,10
+1094,10
@@
def valid_date(in_str: str) -> bool:
import dateparse.dateparse_utils as dp
try:
import dateparse.dateparse_utils as dp
try:
- d = dp.DateParser()
+ d = dp.DateParser()
# type: ignore
_ = d.parse(in_str)
return True
_ = d.parse(in_str)
return True
- except dp.ParseException:
+ except dp.ParseException:
# type: ignore
msg = f'Unable to parse date {in_str}.'
logger.warning(msg)
return False
msg = f'Unable to parse date {in_str}.'
logger.warning(msg)
return False
@@
-1121,9
+1110,9
@@
def to_datetime(in_str: str) -> Optional[datetime.datetime]:
import dateparse.dateparse_utils as dp
try:
import dateparse.dateparse_utils as dp
try:
- d = dp.DateParser()
+ d = dp.DateParser()
# type: ignore
dt = d.parse(in_str)
dt = d.parse(in_str)
- if
type(dt) == datetime.datetime
:
+ if
isinstance(dt, datetime.datetime)
:
return dt
except ValueError:
msg = f'Unable to parse datetime {in_str}.'
return dt
except ValueError:
msg = f'Unable to parse datetime {in_str}.'
@@
-1366,20
+1355,23
@@
def make_contractions(txt: str) -> str:
),
]
),
]
- # Special cases
+ # Special cases
: can't, shan't and won't.
txt = re.sub(r'\b(can)\s*no(t)\b', r"\1'\2", txt, count=0, flags=re.IGNORECASE)
txt = re.sub(r'\b(can)\s*no(t)\b', r"\1'\2", txt, count=0, flags=re.IGNORECASE)
+ txt = re.sub(r'\b(sha)ll\s*(n)o(t)\b', r"\1\2'\3", txt, count=0, flags=re.IGNORECASE)
txt = re.sub(
txt = re.sub(
- r'\b(sha)ll\s*(n)o(t)\b', r"\1\2'\3", txt, count=0, flags=re.IGNORECASE
- )
- txt = re.sub(
- r'\b(w)ill\s*(n)(o)(t)\b', r"\1\3\2'\4", txt, count=0, flags=re.IGNORECASE
+ r'\b(w)ill\s*(n)(o)(t)\b',
+ r"\1\3\2'\4",
+ txt,
+ count=0,
+ flags=re.IGNORECASE,
)
for first_list, second_list in first_second:
for first in first_list:
for second in second_list:
)
for first_list, second_list in first_second:
for first in first_list:
for second in second_list:
- # Disallow there're. It's valid English but sounds weird.
- if first == 'there' and second == 'a(re)':
+ # Disallow there're/where're. They're valid English
+ # but sound weird.
+ if (first in ('there', 'where')) and second == 'a(re)':
continue
pattern = fr'\b({first})\s+{second}\b'
continue
pattern = fr'\b({first})\s+{second}\b'
@@
-1465,11
+1457,11
@@
def shuffle_columns_into_list(
# Column specs map input lines' columns into outputs.
# [col1, col2...]
for spec in column_specs:
# Column specs map input lines' columns into outputs.
# [col1, col2...]
for spec in column_specs:
-
c
hunk = ''
+ hunk = ''
for n in spec:
for n in spec:
-
chunk = c
hunk + delim + input_lines[n]
-
chunk = c
hunk.strip(delim)
- out.append(
c
hunk)
+
hunk =
hunk + delim + input_lines[n]
+
hunk =
hunk.strip(delim)
+ out.append(hunk)
return out
return out
@@
-1495,11
+1487,11
@@
def shuffle_columns_into_dict(
# Column specs map input lines' columns into outputs.
# "key", [col1, col2...]
for spec in column_specs:
# Column specs map input lines' columns into outputs.
# "key", [col1, col2...]
for spec in column_specs:
-
c
hunk = ''
+ hunk = ''
for n in spec[1]:
for n in spec[1]:
-
chunk = c
hunk + delim + input_lines[n]
-
chunk = c
hunk.strip(delim)
- out[spec[0]] =
c
hunk
+
hunk =
hunk + delim + input_lines[n]
+
hunk =
hunk.strip(delim)
+ out[spec[0]] = hunk
return out
return out
@@
-1524,9
+1516,9
@@
def to_ascii(x: str):
b'1, 2, 3'
"""
b'1, 2, 3'
"""
- if
type(x) is str
:
+ if
isinstance(x, str)
:
return x.encode('ascii')
return x.encode('ascii')
- if
type(x) is bytes
:
+ if
isinstance(x, bytes)
:
return x
raise Exception('to_ascii works with strings and bytes')
return x
raise Exception('to_ascii works with strings and bytes')
@@
-1589,9
+1581,7
@@
def chunk(txt: str, chunk_size):
yield txt[x : x + chunk_size]
yield txt[x : x + chunk_size]
-def to_bitstring(
- txt: str, *, delimiter='', encoding='utf-8', errors='surrogatepass'
-) -> str:
+def to_bitstring(txt: str, *, delimiter='', encoding='utf-8', errors='surrogatepass') -> str:
"""Encode txt and then chop it into bytes. Note: only bitstrings
with delimiter='' are interpretable by from_bitstring.
"""Encode txt and then chop it into bytes. Note: only bitstrings
with delimiter='' are interpretable by from_bitstring.