X-Git-Url: https://wannabe.guru.org/gitweb/?a=blobdiff_plain;f=generic_news_rss_renderer.py;h=61be6ff01c487122cf3215a63eda3f31463223c2;hb=477c84bcdefc573278ec8ed8974d5be8baac562d;hp=71cf7ed2d0ceee7191f69d818684e5b65dfc17ab;hpb=e4dca16bbd329afdb587e8488767d88e17777254;p=kiosk.git diff --git a/generic_news_rss_renderer.py b/generic_news_rss_renderer.py index 71cf7ed..61be6ff 100644 --- a/generic_news_rss_renderer.py +++ b/generic_news_rss_renderer.py @@ -4,7 +4,7 @@ from abc import abstractmethod import datetime from dateutil.parser import parse import http.client -import random +import logging import re from typing import Dict, List, Optional, Union import xml.etree.ElementTree as ET @@ -16,7 +16,10 @@ import page_builder import profanity_filter -class generic_news_rss_renderer(renderer.debuggable_abstaining_renderer): +logger = logging.getLogger(__file__) + + +class generic_news_rss_renderer(renderer.abstaining_renderer): def __init__( self, name_to_timeout_dict: Dict[str, int], @@ -24,18 +27,13 @@ class generic_news_rss_renderer(renderer.debuggable_abstaining_renderer): feed_uris: List[str], page_title: str, ): - super(generic_news_rss_renderer, self).__init__(name_to_timeout_dict, False) - self.debug = True + super().__init__(name_to_timeout_dict) self.feed_site = feed_site self.feed_uris = feed_uris self.page_title = page_title self.news = grab_bag.grab_bag() self.details = grab_bag.grab_bag() - self.filter = profanity_filter.profanity_filter() - - @abstractmethod - def debug_prefix(self) -> str: - pass + self.filter = profanity_filter.ProfanityFilter() @abstractmethod def get_headlines_page_prefix(self) -> str: @@ -61,13 +59,17 @@ class generic_news_rss_renderer(renderer.debuggable_abstaining_renderer): def find_title(self, item: ET.Element) -> Optional[str]: return item.findtext("title") - def munge_title(self, title: str) -> str: + def munge_title(self, title: str, item: ET.Element) -> str: return title def find_description(self, item: ET.Element) -> Optional[str]: return item.findtext("description") - def munge_description(self, description: str) -> str: + def munge_description( + self, + description: str, + item: ET.Element + ) -> str: description = re.sub("<[^>]+>", "", description) return description @@ -94,6 +96,12 @@ class generic_news_rss_renderer(renderer.debuggable_abstaining_renderer): ) -> bool: return True + def do_headlines(self) -> bool: + return True + + def do_details(self) -> bool: + return True + def is_item_older_than_n_days(self, item: ET.Element, n: int) -> bool: pubdate = self.find_pubdate(item) if pubdate is None: @@ -118,73 +126,76 @@ class generic_news_rss_renderer(renderer.debuggable_abstaining_renderer): raise Exception def shuffle_news(self) -> bool: - headlines = page_builder.page_builder() - headlines.set_layout(page_builder.page_builder.LAYOUT_FOUR_ITEMS) - headlines.set_title("%s" % self.page_title) - subset = self.news.subset(4) - if subset is None: - self.debug_print("Not enough messages to choose from.") - return False - for msg in subset: - headlines.add_item(msg) - headlines.set_custom_html( - """ -""" - ) - _ = f"{self.get_headlines_page_prefix()}_{self.get_headlines_page_priority()}_25900.html" - with file_writer.file_writer(_) as f: - headlines.render_html(f) - - details = page_builder.page_builder() - details.set_layout(page_builder.page_builder.LAYOUT_ONE_ITEM) - details.set_custom_html( - """ -""" - ) - details.set_title(f"{self.page_title}") - subset = self.details.subset(1) - if subset is None: - self.debug_print("Not enough details to choose from.") - return False - for msg in subset: - blurb = msg - blurb += "" - details.add_item(blurb) - _ = f"{self.get_details_page_prefix()}_{self.get_details_page_priority()}_86400.html" - with file_writer.file_writer(_) as g: - details.render_html(g) + if self.do_headlines(): + headlines = page_builder.page_builder() + headlines.set_layout(page_builder.page_builder.LAYOUT_FOUR_ITEMS) + headlines.set_title("%s" % self.page_title) + subset = self.news.subset(4) + if subset is None: + logger.warning('Not enough messages to select from in shuffle_news?!') + return False + for msg in subset: + headlines.add_item(msg) + headlines.set_custom_html( + """ + """ + ) + _ = f"{self.get_headlines_page_prefix()}_{self.get_headlines_page_priority()}_25900.html" + with file_writer.file_writer(_) as f: + headlines.render_html(f) + + if self.do_details(): + details = page_builder.page_builder() + details.set_layout(page_builder.page_builder.LAYOUT_ONE_ITEM) + details.set_custom_html( + """ + """ + ) + details.set_title(self.page_title) + subset = self.details.subset(1) + if subset is None: + logger.warning('Not enough details to choose from in do_details') + logger.debug("Not enough details to choose from.") + return False + for msg in subset: + blurb = msg + blurb += "" + details.add_item(blurb) + _ = f"{self.get_details_page_prefix()}_{self.get_details_page_priority()}_86400.html" + with file_writer.file_writer(_) as g: + details.render_html(g) return True def fetch_news(self) -> bool: @@ -195,45 +206,57 @@ a:active { http.client.HTTPSConnection]] = None for uri in self.feed_uris: + url = None if self.should_use_https(): - self.debug_print("Fetching: https://%s%s" % (self.feed_site, uri)) - self.conn = http.client.HTTPSConnection(self.feed_site, timeout=20) + url = f'https://{self.feed_site}{uri}' + logger.info(f'Fetching: {url}') + self.conn = http.client.HTTPSConnection(self.feed_site, timeout=10) else: - self.debug_print("Fetching: http://%s%s" % (self.feed_site, uri)) - self.conn = http.client.HTTPConnection(self.feed_site, timeout=20) - assert(self.conn is not None) + url = f'http://{self.feed_site}{uri}' + logger.info(f'Fetching: {url}') + self.conn = http.client.HTTPConnection(self.feed_site, timeout=10) + assert self.conn is not None + assert url is not None self.conn.request( "GET", uri, None, { "Accept": "*/*", - "Cache-control": "max-age=59", - "User-Agent": "Mozilla/5.0 (Macintosh; Intel Mac OS X 10_7_2) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/27.0.1453.93 Safari/537.36", + "Cache-control": "max-age=50", }, ) try: response = self.conn.getresponse() - except: - print("Exception in generic RSS renderer HTTP connection") + except Exception as e: + logger.exception(e) + logger.error( + f"Exception in generic RSS renderer HTTP connection fetching {url}; giving up." + ) return False if response.status != 200: - print( - f"{self.page_title}: RSS fetch_news error, response: {response.status}" + logger.error( + f'Unexpected status {response.status} while fetching {url}; giving up.' ) - self.debug_print(str(response.read())) return False - rss = ET.fromstring(response.read()) + raw = response.read() + logger.info(f'Status 200: got {len(raw)} bytes back from {url}') + rss = ET.fromstring(raw) channel = rss[0] - for item in channel.getchildren(): + title_filter = set() + for item in list(channel): title = self.find_title(item) - if title is not None: - title = self.munge_title(title) description = item.findtext("description") + if title is not None: + title = self.munge_title(title, item) + else: + logger.info('Skipping RSS feed item with no title.') + continue + logger.debug(f'Considering RSS item {title}...') if description is not None: - description = self.munge_description(description) + description = self.munge_description(description, item) else: description = "" image = self.find_image(item) @@ -242,23 +265,27 @@ a:active { link = item.findtext("link") if link is not None: link = self.munge_link(link) - - if title is None or not self.item_is_interesting_for_headlines( - title, description, item + if not self.item_is_interesting_for_headlines( + title, description, item ): - self.debug_print(f'Item "{title}" is not interesting') + logger.info(f'Skipping {title} because it\'s not interesting.') continue if self.should_profanity_filter() and ( - self.filter.contains_bad_words(title) - or self.filter.contains_bad_words(description) + self.filter.contains_bad_word(title) + or self.filter.contains_bad_word(description) ): - self.debug_print(f'Found bad words in item "{title}"') + logger.info(f'Skipping {title} because it contains profanity.') continue + if title in title_filter: + logger.info(f'Skipping {title} because we already saw an item with the same title.') + continue + title_filter.add(title) + blurb = """
""" + font-size:34pt; + -webkit-column-break-inside:avoid;">""" if image is not None: blurb += f'{ts.strftime("%b %d")}' if self.item_is_interesting_for_article(title, description, item): + logger.info(f'Item {title} is also interesting as an article details page; creating...') longblurb = blurb longblurb += "
" longblurb += description longblurb += "
" longblurb = longblurb.replace("font-size:34pt", "font-size:44pt") self.details.add(longblurb) + else: + logger.info(f'Item {title} isn\'t interesting for article details page; skipped.') blurb += "" self.news.add(blurb) count += 1 + logger.debug(f'Added {count} items so far...') return count > 0