__filename__ = "newswire.py" __author__ = "Bob Mottram" __license__ = "AGPL3+" __version__ = "1.5.0" __maintainer__ = "Bob Mottram" __email__ = "bob@libreserver.org" __status__ = "Production" __module_group__ = "Web Interface Columns" import os import json import requests import random import time from socket import error as SocketError import errno from datetime import timedelta from datetime import timezone from collections import OrderedDict from utils import valid_post_date from categories import set_hashtag_category from flags import is_suspended from flags import is_local_network_address from flags import is_public_post from utils import data_dir from utils import string_contains from utils import image_mime_types_dict from utils import resembles_url from utils import get_url_from_post from utils import remove_zero_length_strings from utils import date_from_string_format from utils import acct_handle_dir from utils import remove_eol from utils import get_domain_from_actor from utils import valid_hash_tag from utils import dangerous_svg from utils import get_fav_filename_from_url from utils import get_base_content_from_post from utils import has_object_dict from utils import first_paragraph_from_string from utils import locate_post from utils import load_json from utils import save_json from utils import contains_invalid_chars from utils import remove_html from utils import is_account_dir from utils import acct_dir from utils import local_actor_url from utils import escape_text from utils import unescaped_text from blocking import is_blocked_domain from blocking import is_blocked_hashtag from filters import is_filtered from session import download_image_any_mime_type from content import remove_script def _remove_cdata(text: str) -> str: """Removes any CDATA from the given text """ if 'CDATA[' in text: text = text.split('CDATA[')[1] if ']' in text: text = text.split(']')[0] return text def rss2header(http_prefix: str, nickname: str, domain_full: str, title: str, translate: {}) -> str: """Header for an RSS 2.0 feed """ rss_str = \ "" + \ "" + \ '' if title.startswith('News'): rss_str += \ ' Newswire' + \ ' ' + http_prefix + '://' + domain_full + \ '/newswire.xml' + '' elif title.startswith('Site'): rss_str += \ ' ' + domain_full + '' + \ ' ' + http_prefix + '://' + domain_full + \ '/blog/rss.xml' + '' else: title_str = escape_text(translate[title]) rss_str += \ ' ' + title_str + '' + \ ' ' + \ local_actor_url(http_prefix, nickname, domain_full) + \ '/rss.xml' + '' return rss_str def rss2footer() -> str: """Footer for an RSS 2.0 feed """ rss_str = '' return rss_str def get_newswire_tags(text: str, max_tags: int) -> []: """Returns a list of hashtags found in the given text """ if '#' not in text: return [] if ' ' not in text: return [] text_simplified = \ text.replace(',', ' ').replace(';', ' ').replace('- ', ' ') text_simplified = text_simplified.replace('. ', ' ').strip() if text_simplified.endswith('.'): text_simplified = text_simplified[:len(text_simplified)-1] words = text_simplified.split(' ') tags = [] for wrd in words: if not wrd.startswith('#'): continue if len(wrd) <= 1: continue if wrd in tags: continue tags.append(wrd) if len(tags) >= max_tags: break return tags def limit_word_lengths(text: str, max_word_length: int) -> str: """Limits the maximum length of words so that the newswire column cannot become too wide """ if ' ' not in text: return text words = text.split(' ') result = '' for wrd in words: if len(wrd) > max_word_length: wrd = wrd[:max_word_length] if result: result += ' ' result += wrd return result def get_newswire_favicon_url(url: str) -> str: """Returns a favicon url from the given article link """ if '://' not in url: return '/newswire_favicon.ico' if url.startswith('http://'): if not (url.endswith('.onion') or url.endswith('.i2p')): return '/newswire_favicon.ico' domain = url.split('://')[1] if '/' not in domain: return url + '/favicon.ico' domain = domain.split('/')[0] return url.split('://')[0] + '://' + domain + '/favicon.ico' def _download_newswire_feed_favicon(session, base_dir: str, link: str, debug: bool) -> bool: """Downloads the favicon for the given feed link """ fav_url = get_newswire_favicon_url(link) if '://' not in link: return False timeout_sec = 10 image_data, mime_type = \ download_image_any_mime_type(session, fav_url, timeout_sec, debug) if not image_data or not mime_type: return False # update the favicon url extensions_to_mime = image_mime_types_dict() for ext, mime_ext in extensions_to_mime.items(): if 'image/' + mime_ext in mime_type: fav_url = fav_url.replace('.ico', '.' + ext) break # create cached favicons directory if needed if not os.path.isdir(base_dir + '/favicons'): os.mkdir(base_dir + '/favicons') # check svg for dubious scripts if fav_url.endswith('.svg'): image_data_str = str(image_data) if dangerous_svg(image_data_str, False): return False # save to the cache fav_filename = get_fav_filename_from_url(base_dir, fav_url) if os.path.isfile(fav_filename): return True try: with open(fav_filename, 'wb+') as fp_fav: fp_fav.write(image_data) except OSError: print('EX: failed writing favicon ' + fav_filename) return False return True def _add_newswire_dict_entry(base_dir: str, newswire: {}, date_str: str, title: str, link: str, votes_status: str, post_filename: str, description: str, moderated: bool, mirrored: bool, tags: [], max_tags: int, session, debug: bool, podcast_properties: {}, system_language: str, fediverse_handle: str, extra_links: []) -> None: """Update the newswire dictionary """ # remove any markup title = remove_html(title) description = remove_html(description) all_text = title + ' ' + description # check that none of the text is filtered against if is_filtered(base_dir, None, None, all_text, system_language): return title = limit_word_lengths(title, 13) if tags is None: tags = [] # extract hashtags from the text of the feed post post_tags = get_newswire_tags(all_text, max_tags) # Include tags from podcast categories if podcast_properties: if podcast_properties.get('explicit'): if '#nsfw' not in post_tags: post_tags.append('#nsfw') post_tags += podcast_properties['categories'] # combine the tags into a single list for tag in tags: if tag in post_tags: continue if len(post_tags) < max_tags: post_tags.append(tag) # check that no tags are blocked for tag in post_tags: if is_blocked_hashtag(base_dir, tag): return _download_newswire_feed_favicon(session, base_dir, link, debug) newswire[date_str] = [ title, link, votes_status, post_filename, description, moderated, post_tags, mirrored, podcast_properties, fediverse_handle, extra_links ] def _valid_feed_date(pub_date: str, debug: bool = False) -> bool: # convert from YY-MM-DD HH:MM:SS+00:00 to # YY-MM-DDTHH:MM:SSZ post_date = pub_date.replace(' ', 'T').replace('+00:00', 'Z') if '.' in post_date: ending = post_date.split('.')[1] timezone_str = '' for ending_char in ending: if not ending_char.isdigit(): timezone_str += ending_char if timezone_str: post_date = post_date.split('.')[0] + timezone_str return valid_post_date(post_date, 90, debug) def parse_feed_date(pub_date: str, unique_string_identifier: str) -> str: """Returns a UTC date string based on the given date string This tries a number of formats to see which work """ if ':00:00' in pub_date: # If this was published exactly on the hour then assign a # random minute and second to make this item relatively unique randgen = random.Random(unique_string_identifier) rand_min = randgen.randint(0, 59) rand_sec = randgen.randint(0, 59) replace_time_str = \ ':' + str(rand_min).zfill(2) + ':' + str(rand_sec).zfill(2) pub_date = pub_date.replace(':00:00', replace_time_str) formats = ("%a, %d %b %Y %H:%M:%S %z", "%a, %d %b %Y %H:%M:%S Z", "%a, %d %b %Y %H:%M:%S GMT", "%a, %d %b %Y %H:%M:%S EST", "%a, %d %b %Y %H:%M:%S PST", "%a, %d %b %Y %H:%M:%S AST", "%a, %d %b %Y %H:%M:%S CST", "%a, %d %b %Y %H:%M:%S MST", "%a, %d %b %Y %H:%M:%S AKST", "%a, %d %b %Y %H:%M:%S HST", "%a, %d %b %Y %H:%M:%S UT", "%Y-%m-%dT%H:%M:%S%z", "%Y-%m-%dT%H:%M:%S%Z") published_date = None for date_format in formats: if ',' in pub_date and ',' not in date_format: continue if ',' not in pub_date and ',' in date_format: continue if 'Z' in pub_date and 'Z' not in date_format: continue if 'Z' not in pub_date and 'Z' in date_format: continue if 'EST' not in pub_date and 'EST' in date_format: continue if 'GMT' not in pub_date and 'GMT' in date_format: continue if 'EST' in pub_date and 'EST' not in date_format: continue if 'UT' not in pub_date and 'UT' in date_format: continue if 'UT' in pub_date and 'UT' not in date_format: continue # remove any fraction of a second pub_date2 = pub_date if '.' in pub_date2: ending = pub_date2.split('.')[1] timezone_str = '' if '+' in ending: timezone_str = '+' + ending.split('+')[1] elif '-' in ending: timezone_str = '-' + ending.split('-')[1] pub_date2 = pub_date2.split('.')[0] + timezone_str try: published_date = date_from_string_format(pub_date2, [date_format]) except BaseException: continue if published_date: if pub_date.endswith(' EST'): hours_added = timedelta(hours=5) published_date = published_date + hours_added break pub_date_str = None if published_date: offset = published_date.utcoffset() if offset: published_date = published_date - offset # convert local date to UTC published_date = published_date.replace(tzinfo=timezone.utc) pub_date_str = str(published_date) if not pub_date_str.endswith('+00:00'): pub_date_str += '+00:00' else: print('WARN: unrecognized date format: ' + pub_date) return pub_date_str def load_hashtag_categories(base_dir: str, language: str) -> None: """Loads an rss file containing hashtag categories """ hashtag_categories_filename = base_dir + '/categories.xml' if not os.path.isfile(hashtag_categories_filename): hashtag_categories_filename = \ base_dir + '/defaultcategories/' + language + '.xml' if not os.path.isfile(hashtag_categories_filename): return try: with open(hashtag_categories_filename, 'r', encoding='utf-8') as fp_cat: xml_str = fp_cat.read() _xml2str_to_hashtag_categories(base_dir, xml_str, 1024, True) except OSError: print('EX: load_hashtag_categories unable to read ' + hashtag_categories_filename) def _xml2str_to_hashtag_categories(base_dir: str, xml_str: str, max_categories_feed_item_size_kb: int, force: bool = False) -> None: """Updates hashtag categories based upon an rss feed """ rss_items = xml_str.split('') max_bytes = max_categories_feed_item_size_kb * 1024 for rss_item in rss_items: if not rss_item: continue if len(rss_item) > max_bytes: print('WARN: rss categories feed item is too big') continue if '' not in rss_item: continue if '' not in rss_item: continue if '' not in rss_item: continue if '' not in rss_item: continue category_str = rss_item.split('')[1] category_str = category_str.split('')[0].strip() category_str = unescaped_text(category_str) if not category_str: continue if 'CDATA' in category_str: continue hashtag_list_str = rss_item.split('')[1] hashtag_list_str = hashtag_list_str.split('')[0].strip() hashtag_list_str = unescaped_text(hashtag_list_str) if not hashtag_list_str: continue if 'CDATA' in hashtag_list_str: continue hashtag_list = hashtag_list_str.split(' ') if is_blocked_hashtag(base_dir, category_str): continue for hashtag in hashtag_list: set_hashtag_category(base_dir, hashtag, category_str, False, force) def _get_podcast_categories(xml_item: str, xml_str: str) -> str: """ get podcast categories if they exist. These can be turned into hashtags See https://podcast-standard.org/itunes_tags """ podcast_categories = [] # convert keywords to hashtags if '' in keywords_str: keywords_str = keywords_str.split('>')[1] if '<' in keywords_str: keywords_str = keywords_str.split('<')[0] keywords_str = remove_html(keywords_str) keywords_list = keywords_str.split(',') for keyword in keywords_list: keyword_hashtag = '#' + keyword.strip() if keyword_hashtag not in podcast_categories: if valid_hash_tag(keyword): podcast_categories.append(keyword_hashtag) episode_category_tags = ['' in episode_category: episode_category = episode_category.split('>')[1] if '<' in episode_category: episode_category = episode_category.split('<')[0] episode_category = \ episode_category.lower().replace(' ', '') episode_category = episode_category.replace('#', '') episode_category_hashtag = '#' + episode_category if episode_category_hashtag not in podcast_categories: if valid_hash_tag(episode_category): podcast_categories.append(episode_category_hashtag) return podcast_categories def _get_podcast_author(xml_item: str, xml_str: str) -> str: """ get podcast author if specified. """ author = None episode_author_tags = ['' not in author_str: continue author_str = author_str.split('>')[1] if '<' not in author_str: continue author = item_str.split('>')[0] return remove_html(author).strip() return author def _valid_podcast_entry(base_dir: str, key: str, entry: {}) -> bool: """Is the given podcast namespace entry valid? https://github.com/Podcastindex-org/podcast-namespace/ blob/main/proposal-docs/social/social.md#socialinteract-element """ if key in ('socialInteract', 'discussion'): if not entry.get('protocol'): return False if not entry.get('uri'): if not entry.get('text'): if not entry.get('url'): return False if entry['protocol'].tolower() != 'activitypub': return False if entry.get('uri'): post_url = remove_html(entry['uri']) elif entry.get('url'): post_url = remove_html(entry['uri']) else: post_url = entry['text'] if '://' not in post_url: return False post_domain, _ = get_domain_from_actor(post_url) if not post_domain: return False if is_blocked_domain(base_dir, post_domain, None, None): return False return True def xml_podcast_to_dict(base_dir: str, xml_item: str, xml_str: str) -> {}: """podcasting extensions for RSS feeds See https://github.com/Podcastindex-org/podcast-namespace/ blob/main/docs/1.0.md https://github.com/Podcastindex-org/podcast-namespace/ blob/main/proposal-docs/social/social.md#socialinteract-element """ if '' not in pod_line: ctr += 1 continue if ' ' not in pod_line.split('>')[0]: pod_key = pod_line.split('>')[0].strip() pod_val = pod_line.split('>', 1)[1].strip() if '<' in pod_val: pod_val = pod_val.split('<')[0] if pod_key in podcast_properties: podcast_properties[pod_key] = pod_val ctr += 1 continue pod_key = pod_line.split(' ')[0] pod_fields = ( 'url', 'geo', 'osm', 'type', 'method', 'group', 'owner', 'srcset', 'img', 'role', 'address', 'suggested', 'startTime', 'duration', 'href', 'name', 'pubdate', 'length', 'season', 'email', 'platform', 'protocol', 'accountId', 'priority', 'podcastAccountId', 'podcastAccountUrl' ) pod_entry = {} for pod_field in pod_fields: if pod_field + '="' not in pod_line: continue pod_str = pod_line.split(pod_field + '="')[1] if '"' not in pod_str: continue pod_val = pod_str.split('"')[0] pod_entry[pod_field] = pod_val pod_text = pod_line.split('>')[1] if '<' in pod_text: pod_text = pod_text.split('<')[0].strip() if pod_text: pod_entry['text'] = pod_text appended = False if pod_key + 's' in podcast_properties: if isinstance(podcast_properties[pod_key + 's'], list): podcast_properties[pod_key + 's'].append(pod_entry) appended = True if not appended: # if there are repeated keys then only use the first one if not podcast_properties.get(pod_key): if _valid_podcast_entry(base_dir, pod_key, pod_entry): podcast_properties[pod_key] = pod_entry ctr += 1 # get the image for the podcast, if it exists podcast_episode_image = None episode_image_tags = ['' in episode_image: episode_image = episode_image.split('>')[0] if 'href="' in episode_image: episode_image = episode_image.split('href="')[1] if '"' in episode_image: episode_image = episode_image.split('"')[0] podcast_episode_image = episode_image break elif 'url="' in episode_image: episode_image = episode_image.split('url="')[1] if '"' in episode_image: episode_image = episode_image.split('"')[0] podcast_episode_image = episode_image break elif '>' in episode_image: episode_image = episode_image.split('>')[1] if '<' in episode_image: episode_image = episode_image.split('<')[0] if resembles_url(episode_image): podcast_episode_image = episode_image break # get categories if they exist. These can be turned into hashtags podcast_categories = _get_podcast_categories(xml_item, xml_str) # get the author name podcast_author = _get_podcast_author(xml_item, xml_str) if podcast_author: podcast_properties['author'] = podcast_author if podcast_episode_image: podcast_properties['image'] = podcast_episode_image podcast_properties['categories'] = podcast_categories if string_contains(xml_item, ('Y', 'T', '1')): podcast_properties['explicit'] = True else: podcast_properties['explicit'] = False else: if ' (str, str): """Extracts rss link from rss item string """ mime_type = None if preferred_mime_types and '' in enclosure: enclosure = enclosure.split('>')[0] if ' type="' in enclosure: mime_type = enclosure.split(' type="')[1] if '"' in mime_type: mime_type = mime_type.split('"')[0] if 'url="' in enclosure and \ ('"audio/' in enclosure or '"video/' in enclosure): link_str = enclosure.split('url="')[1] if '"' in link_str: link = link_str.split('"')[0] if resembles_url(link): return link, mime_type if '' in rss_item and '' in rss_item: link = rss_item.split('')[1] link = link.split('')[0] if '://' not in link: return None, None elif '' in link_str: link_str = link_str.split('>')[0] if 'href="' in link_str: link_str = link_str.split('href="')[1] if '"' in link_str: link = link_str.split('"')[0] return link, mime_type def _xml2str_to_dict(base_dir: str, domain: str, xml_str: str, moderated: bool, mirrored: bool, max_posts_per_source: int, max_feed_item_size_kb: int, max_categories_feed_item_size_kb: int, session, debug: bool, preferred_podcast_formats: [], system_language: str) -> {}: """Converts an xml RSS 2.0 string to a dictionary """ if '' not in xml_str: return {} result = {} # is this an rss feed containing hashtag categories? if '#categories' in xml_str: _xml2str_to_hashtag_categories(base_dir, xml_str, max_categories_feed_item_size_kb) return {} rss_items = xml_str.split('') post_ctr = 0 max_bytes = max_feed_item_size_kb * 1024 first_item = True for rss_item in rss_items: if first_item: first_item = False continue if not rss_item: continue if len(rss_item) > max_bytes: print('WARN: rss feed item is too big') continue if '' not in rss_item: continue if '' not in rss_item: continue if '' not in rss_item: continue if '' not in rss_item: continue title = rss_item.split('')[1] title = _remove_cdata(title.split('')[0]) title = unescaped_text(title) title = remove_script(title, None, None, None) title = remove_html(title) title = title.replace('\n', '') description = '' if '' in rss_item and '' in rss_item: description = rss_item.split('')[1] description = description.split('')[0] description = unescaped_text(description) description = remove_script(description, None, None, None) description = remove_html(description) else: if '' in rss_item and \ '' in rss_item: description = rss_item.split('')[1] description = description.split('')[0] description = unescaped_text(description) description = remove_script(description, None, None, None) description = remove_html(description) proxy_type = None if domain.endswith('.onion'): proxy_type = 'tor' elif domain.endswith('.i2p'): proxy_type = 'i2p' link, link_mime_type = \ get_link_from_rss_item(rss_item, preferred_podcast_formats, proxy_type) if not link: continue item_domain = link.split('://')[1] if '/' in item_domain: item_domain = item_domain.split('/')[0] if is_blocked_domain(base_dir, item_domain, None, None): continue pub_date = rss_item.split('')[1] pub_date = pub_date.split('')[0] unique_string_identifier = title + ' ' + link pub_date_str = parse_feed_date(pub_date, unique_string_identifier) if not pub_date_str: continue if not _valid_feed_date(pub_date_str): continue post_filename = '' votes_status = [] podcast_properties = \ xml_podcast_to_dict(base_dir, rss_item, xml_str) if podcast_properties: podcast_properties['linkMimeType'] = link_mime_type fediverse_handle = '' extra_links = [] _add_newswire_dict_entry(base_dir, result, pub_date_str, title, link, votes_status, post_filename, description, moderated, mirrored, [], 32, session, debug, podcast_properties, system_language, fediverse_handle, extra_links) post_ctr += 1 if post_ctr >= max_posts_per_source: break if post_ctr > 0: print('Added ' + str(post_ctr) + ' rss 2.0 feed items to newswire') return result def _xml1str_to_dict(base_dir: str, domain: str, xml_str: str, moderated: bool, mirrored: bool, max_posts_per_source: int, max_feed_item_size_kb: int, max_categories_feed_item_size_kb: int, session, debug: bool, preferred_podcast_formats: [], system_language: str) -> {}: """Converts an xml RSS 1.0 string to a dictionary https://validator.w3.org/feed/docs/rss1.html """ item_str = '#categories' in xml_str: _xml2str_to_hashtag_categories(base_dir, xml_str, max_categories_feed_item_size_kb) return {} rss_items = xml_str.split(item_str) post_ctr = 0 max_bytes = max_feed_item_size_kb * 1024 first_item = True for rss_item in rss_items: if first_item: first_item = False continue if not rss_item: continue if len(rss_item) > max_bytes: print('WARN: rss 1.0 feed item is too big') continue if rss_item.startswith('s>'): continue if '' not in rss_item: continue if '' not in rss_item: continue if '' not in rss_item: continue if '' not in rss_item: continue title = rss_item.split('')[1] title = _remove_cdata(title.split('')[0]) title = unescaped_text(title) title = remove_script(title, None, None, None) title = remove_html(title) description = '' if '' in rss_item and '' in rss_item: description = rss_item.split('')[1] description = description.split('')[0] description = unescaped_text(description) description = remove_script(description, None, None, None) description = remove_html(description) else: if '' in rss_item and \ '' in rss_item: description = rss_item.split('')[1] description = description.split('')[0] description = unescaped_text(description) description = remove_script(description, None, None, None) description = remove_html(description) proxy_type = None if domain.endswith('.onion'): proxy_type = 'tor' elif domain.endswith('.i2p'): proxy_type = 'i2p' link, link_mime_type = \ get_link_from_rss_item(rss_item, preferred_podcast_formats, proxy_type) if not link: continue item_domain = link.split('://')[1] if '/' in item_domain: item_domain = item_domain.split('/')[0] if is_blocked_domain(base_dir, item_domain, None, None): continue pub_date = rss_item.split('')[1] pub_date = pub_date.split('')[0] unique_string_identifier = title + ' ' + link pub_date_str = parse_feed_date(pub_date, unique_string_identifier) if not pub_date_str: continue if not _valid_feed_date(pub_date_str): continue post_filename = '' votes_status = [] podcast_properties = \ xml_podcast_to_dict(base_dir, rss_item, xml_str) if podcast_properties: podcast_properties['linkMimeType'] = link_mime_type fediverse_handle = '' extra_links = [] _add_newswire_dict_entry(base_dir, result, pub_date_str, title, link, votes_status, post_filename, description, moderated, mirrored, [], 32, session, debug, podcast_properties, system_language, fediverse_handle, extra_links) post_ctr += 1 if post_ctr >= max_posts_per_source: break if post_ctr > 0: print('Added ' + str(post_ctr) + ' rss 1.0 feed items to newswire') return result def _atom_feed_to_dict(base_dir: str, domain: str, xml_str: str, moderated: bool, mirrored: bool, max_posts_per_source: int, max_feed_item_size_kb: int, session, debug: bool, preferred_podcast_formats: [], system_language: str) -> {}: """Converts an atom feed string to a dictionary Also see https://activitystrea.ms/specs/atom/1.0/ """ if '' not in xml_str: return {} result = {} atom_items = xml_str.split('') post_ctr = 0 max_bytes = max_feed_item_size_kb * 1024 first_item = True for atom_item in atom_items: if first_item: first_item = False continue if not atom_item: continue if len(atom_item) > max_bytes: print('WARN: atom feed item is too big') continue if '' not in atom_item: continue if '' not in atom_item: continue if '' not in atom_item: continue if '' not in atom_item: continue title = atom_item.split('')[1] title = _remove_cdata(title.split('')[0]) title = unescaped_text(title) title = remove_script(title, None, None, None) title = remove_html(title) description = '' if '' in atom_item and '' in atom_item: description = atom_item.split('')[1] description = unescaped_text(description.split('')[0]) description = remove_script(description, None, None, None) description = remove_html(description) elif '' in atom_item: description = atom_item.split('', 1)[1] description = unescaped_text(description.split('')[0]) description = remove_script(description, None, None, None) description = remove_html(description) else: if '' in atom_item and \ '' in atom_item: description = atom_item.split('')[1] description = description.split('')[0] description = unescaped_text(description) description = remove_script(description, None, None, None) description = remove_html(description) # is there a fediverse handle fediverse_handle = '' if '' in atom_item and '' in atom_item: actor_str = atom_item.split('')[1] actor_str = unescaped_text(actor_str.split('')[0]) actor_str = remove_script(actor_str, None, None, None) if '' in actor_str and \ '' in actor_str and \ '' in actor_str and '' in actor_str: obj_type = actor_str.split('')[1] obj_type = obj_type.split('')[0] if obj_type == 'Person': actor_uri = actor_str.split('')[1] actor_uri = actor_uri.split('')[0] if resembles_url(actor_uri) and \ not is_local_network_address(actor_uri): fediverse_handle = actor_uri # are there any extra links? extra_links = [] if '' in atom_item and \ '' in atom_item: obj_str = atom_item.split('')[1] obj_str = \ unescaped_text(obj_str.split('')[0]) obj_str = remove_script(obj_str, None, None, None) sections = obj_str.split('' in section_str: link_str = section_str.split('>')[0] if 'href="' in link_str and \ 'rel="preview"' not in link_str: link_str = link_str.split('href="')[1] if '"' in link_str: link_str = link_str.split('"')[0] link_str = remove_html(link_str) if resembles_url(link_str) and \ not is_local_network_address(link_str): if link_str not in extra_links: extra_links.append(link_str) proxy_type = None if domain.endswith('.onion'): proxy_type = 'tor' elif domain.endswith('.i2p'): proxy_type = 'i2p' link, link_mime_type = \ get_link_from_rss_item(atom_item, preferred_podcast_formats, proxy_type) if not link: continue item_domain = link.split('://')[1] if '/' in item_domain: item_domain = item_domain.split('/')[0] if is_blocked_domain(base_dir, item_domain, None, None): continue pub_date = atom_item.split('')[1] pub_date = pub_date.split('')[0] unique_string_identifier = title + ' ' + link pub_date_str = parse_feed_date(pub_date, unique_string_identifier) if not pub_date_str: continue if not _valid_feed_date(pub_date_str): continue post_filename = '' votes_status = [] podcast_properties = \ xml_podcast_to_dict(base_dir, atom_item, xml_str) if podcast_properties: podcast_properties['linkMimeType'] = link_mime_type _add_newswire_dict_entry(base_dir, result, pub_date_str, title, link, votes_status, post_filename, description, moderated, mirrored, [], 32, session, debug, podcast_properties, system_language, fediverse_handle, extra_links) post_ctr += 1 if post_ctr >= max_posts_per_source: break if post_ctr > 0: print('Added ' + str(post_ctr) + ' atom feed items to newswire') return result def _json_feed_v1to_dict(base_dir: str, xml_str: str, moderated: bool, mirrored: bool, max_posts_per_source: int, max_feed_item_size_kb: int, session, debug: bool, system_language: str) -> {}: """Converts a json feed string to a dictionary See https://jsonfeed.org/version/1.1 """ if '"items"' not in xml_str: return {} try: feed_json = json.loads(xml_str) except BaseException: print('EX: _json_feed_v1to_dict unable to load json ' + str(xml_str)) return {} max_bytes = max_feed_item_size_kb * 1024 if not feed_json.get('version'): return {} if not feed_json['version'].startswith('https://jsonfeed.org/version/1'): return {} if not feed_json.get('items'): return {} if not isinstance(feed_json['items'], list): return {} post_ctr = 0 result = {} for json_feed_item in feed_json['items']: if not json_feed_item: continue if not isinstance(json_feed_item, dict): continue if not json_feed_item.get('url'): continue url_str = get_url_from_post(json_feed_item['url']) if not url_str: continue if not json_feed_item.get('date_published'): if not json_feed_item.get('date_modified'): continue if not json_feed_item.get('content_text'): if not json_feed_item.get('content_html'): continue if json_feed_item.get('content_html'): if not isinstance(json_feed_item['content_html'], str): continue title = remove_html(json_feed_item['content_html']) else: if not isinstance(json_feed_item['content_text'], str): continue title = remove_html(json_feed_item['content_text']) if len(title) > max_bytes: print('WARN: json feed title is too long') continue description = '' if json_feed_item.get('description'): if not isinstance(json_feed_item['description'], str): continue description = remove_html(json_feed_item['description']) if len(description) > max_bytes: print('WARN: json feed description is too long') continue if json_feed_item.get('tags'): if isinstance(json_feed_item['tags'], list): for tag_name in json_feed_item['tags']: if not isinstance(tag_name, str): continue if ' ' in tag_name: continue if not tag_name.startswith('#'): tag_name = '#' + tag_name if tag_name not in description: description += ' ' + tag_name link = remove_html(url_str) if '://' not in link: continue if len(link) > max_bytes: print('WARN: json feed link is too long') continue item_domain = link.split('://')[1] if '/' in item_domain: item_domain = item_domain.split('/')[0] if is_blocked_domain(base_dir, item_domain, None, None): continue if json_feed_item.get('date_published'): if not isinstance(json_feed_item['date_published'], str): continue pub_date = json_feed_item['date_published'] else: if not isinstance(json_feed_item['date_modified'], str): continue pub_date = json_feed_item['date_modified'] unique_string_identifier = title + ' ' + link pub_date_str = parse_feed_date(pub_date, unique_string_identifier) if not pub_date_str: continue if not _valid_feed_date(pub_date_str): continue post_filename = '' votes_status = [] fediverse_handle = '' extra_links = [] _add_newswire_dict_entry(base_dir, result, pub_date_str, title, link, votes_status, post_filename, description, moderated, mirrored, [], 32, session, debug, None, system_language, fediverse_handle, extra_links) post_ctr += 1 if post_ctr >= max_posts_per_source: break if post_ctr > 0: print('Added ' + str(post_ctr) + ' json feed items to newswire') return result def _atom_feed_yt_to_dict(base_dir: str, xml_str: str, moderated: bool, mirrored: bool, max_posts_per_source: int, max_feed_item_size_kb: int, session, debug: bool, system_language: str) -> {}: """Converts an atom-style YouTube feed string to a dictionary """ if '' not in xml_str: return {} if is_blocked_domain(base_dir, 'www.youtube.com', None, None): return {} result = {} atom_items = xml_str.split('') post_ctr = 0 max_bytes = max_feed_item_size_kb * 1024 first_entry = True for atom_item in atom_items: if first_entry: first_entry = False continue if not atom_item: continue if not atom_item.strip(): continue if len(atom_item) > max_bytes: print('WARN: atom feed item is too big') continue if '' not in atom_item: continue if '' not in atom_item: continue if '' not in atom_item: continue if '' not in atom_item: continue if '' not in atom_item: continue if '' not in atom_item: continue title = atom_item.split('')[1] title = _remove_cdata(title.split('')[0]) title = remove_script(title, None, None, None) title = unescaped_text(title) description = '' if '' in atom_item and \ '' in atom_item: description = atom_item.split('')[1] description = description.split('')[0] description = unescaped_text(description) description = remove_script(description, None, None, None) description = remove_html(description) elif '' in atom_item and '' in atom_item: description = atom_item.split('')[1] description = description.split('')[0] description = unescaped_text(description) description = remove_script(description, None, None, None) description = remove_html(description) elif '' in atom_item: description = atom_item.split('', 1)[1] description = description.split('')[0] description = unescaped_text(description) description = remove_script(description, None, None, None) description = remove_html(description) link, _ = get_link_from_rss_item(atom_item, None, None) if not link: link = atom_item.split('')[1] link = link.split('')[0] link = 'https://www.youtube.com/watch?v=' + link.strip() if not link: continue pub_date = atom_item.split('')[1] pub_date = pub_date.split('')[0] unique_string_identifier = title + ' ' + link pub_date_str = parse_feed_date(pub_date, unique_string_identifier) if not pub_date_str: continue if not _valid_feed_date(pub_date_str): continue post_filename = '' votes_status = [] podcast_properties = \ xml_podcast_to_dict(base_dir, atom_item, xml_str) if podcast_properties: podcast_properties['linkMimeType'] = 'video/youtube' fediverse_handle = '' extra_links = [] _add_newswire_dict_entry(base_dir, result, pub_date_str, title, link, votes_status, post_filename, description, moderated, mirrored, [], 32, session, debug, podcast_properties, system_language, fediverse_handle, extra_links) post_ctr += 1 if post_ctr >= max_posts_per_source: break if post_ctr > 0: print('Added ' + str(post_ctr) + ' YouTube feed items to newswire') return result def _xml_str_to_dict(base_dir: str, domain: str, xml_str: str, moderated: bool, mirrored: bool, max_posts_per_source: int, max_feed_item_size_kb: int, max_categories_feed_item_size_kb: int, session, debug: bool, preferred_podcast_formats: [], system_language: str) -> {}: """Converts an xml string to a dictionary """ if '' in xml_str and '' in xml_str: print('YouTube feed: reading') return _atom_feed_yt_to_dict(base_dir, xml_str, moderated, mirrored, max_posts_per_source, max_feed_item_size_kb, session, debug, system_language) if 'rss version="2.0"' in xml_str: return _xml2str_to_dict(base_dir, domain, xml_str, moderated, mirrored, max_posts_per_source, max_feed_item_size_kb, max_categories_feed_item_size_kb, session, debug, preferred_podcast_formats, system_language) if ' str: """Converts a YouTube channel url into an atom feed url """ if 'youtube.com/channel/' not in url: return url channel_id = url.split('youtube.com/channel/')[1].strip() channel_url = \ 'https://www.youtube.com/feeds/videos.xml?channel_id=' + channel_id print('YouTube feed: ' + channel_url) return channel_url def get_rss(base_dir: str, domain: str, session, url: str, moderated: bool, mirrored: bool, max_posts_per_source: int, max_feed_size_kb: int, max_feed_item_size_kb: int, max_categories_feed_item_size_kb: int, debug: bool, preferred_podcast_formats: [], timeout_sec: int, system_language: str) -> {}: """Returns an RSS url as a dict """ if not isinstance(url, str): print('url: ' + str(url)) print('ERROR: get_rss url should be a string') return None headers = { 'Accept': 'text/xml, application/xml; charset=UTF-8' } params = None session_params = {} session_headers = {} if headers: session_headers = headers if params: session_params = params session_headers['User-Agent'] = \ 'Mozilla/5.0 (X11; Linux x86_64; rv:81.0) Gecko/20100101 Firefox/81.0' if not session: print('WARN: no session specified for get_rss') url = _yt_channel_to_atom_feed(url) try: result = \ session.get(url, headers=session_headers, params=session_params, timeout=timeout_sec, allow_redirects=True) if result: result_str = remove_zero_length_strings(result.text) if int(len(result_str) / 1024) >= max_feed_size_kb: print('WARN: feed is too large: ' + url) elif not contains_invalid_chars(result_str): return _xml_str_to_dict(base_dir, domain, result_str, moderated, mirrored, max_posts_per_source, max_feed_item_size_kb, max_categories_feed_item_size_kb, session, debug, preferred_podcast_formats, system_language) print('WARN: feed contains invalid characters: ' + url) else: print('WARN: no result returned for feed ' + url) except requests.exceptions.RequestException as ex: print('WARN: get_rss failed\nurl: ' + str(url) + ', ' + 'headers: ' + str(session_headers) + ', ' + 'params: ' + str(session_params) + ', ' + str(ex)) except ValueError as ex: print('WARN: get_rss failed\nurl: ' + str(url) + ', ' + 'headers: ' + str(session_headers) + ', ' + 'params: ' + str(session_params) + ', ' + str(ex)) except SocketError as ex: if ex.errno == errno.ECONNRESET: print('WARN: connection was reset during get_rss ' + str(ex)) else: print('WARN: get_rss, ' + str(ex)) return None def get_rss_from_dict(newswire: {}, http_prefix: str, domain_full: str, translate: {}) -> str: """Returns an rss feed from the current newswire dict. This allows other instances to subscribe to the same newswire """ rss_str = rss2header(http_prefix, None, domain_full, 'Newswire', translate) if not newswire: return '' for published, fields in newswire.items(): if '+00:00' in published: published = published.replace('+00:00', 'Z').strip() published = published.replace(' ', 'T') else: published_with_offset = \ date_from_string_format(published, ["%Y-%m-%d %H:%M:%S%z"]) published = published_with_offset.strftime("%Y-%m-%dT%H:%M:%S%z") try: pub_date = date_from_string_format(published, ["%Y-%m-%dT%H:%M:%S%z"]) except BaseException as ex: print('WARN: Unable to convert date ' + published + ' ' + str(ex)) continue rss_str += \ '\n' + \ ' ' + escape_text(fields[0]) + '\n' description = remove_html(first_paragraph_from_string(fields[4])) rss_str += \ ' ' + escape_text(description) + '\n' url = fields[1] if '://' not in url: if domain_full not in url: url = http_prefix + '://' + domain_full + url rss_str += ' ' + url + '\n' rss_date_str = pub_date.strftime("%a, %d %b %Y %H:%M:%S UT") rss_str += \ ' ' + rss_date_str + '\n' + \ '\n' rss_str += rss2footer() return rss_str def _is_newswire_blog_post(post_json_object: {}) -> bool: """Is the given object a blog post? There isn't any difference between a blog post and a newswire blog post but we may here need to check for different properties than is_blog_post does """ if not post_json_object: return False if not has_object_dict(post_json_object): return False if post_json_object['object'].get('summary') and \ post_json_object['object'].get('url') and \ post_json_object['object'].get('content') and \ post_json_object['object'].get('published'): return is_public_post(post_json_object) return False def _get_hashtags_from_post(post_json_object: {}) -> []: """Returns a list of any hashtags within a post """ if not has_object_dict(post_json_object): return [] if not post_json_object['object'].get('tag'): return [] if not isinstance(post_json_object['object']['tag'], list): return [] tags = [] for tgname in post_json_object['object']['tag']: if not isinstance(tgname, dict): continue if not tgname.get('name'): continue if not tgname.get('type'): continue if tgname['type'] != 'Hashtag': continue if tgname['name'] not in tags: tags.append(tgname['name']) return tags def _add_account_blogs_to_newswire(base_dir: str, nickname: str, domain: str, newswire: {}, max_blogs_per_account: int, index_filename: str, max_tags: int, system_language: str, session, debug: bool) -> None: """Adds blogs for the given account to the newswire """ if not os.path.isfile(index_filename): return # local blog entries are unmoderated by default moderated = False # local blogs can potentially be moderated moderated_filename = \ acct_dir(base_dir, nickname, domain) + '/.newswiremoderated' if os.path.isfile(moderated_filename): moderated = True try: with open(index_filename, 'r', encoding='utf-8') as fp_index: post_filename = 'start' ctr = 0 while post_filename: post_filename = fp_index.readline() if not post_filename: ctr += 1 if ctr >= max_blogs_per_account: break continue # if this is a full path then remove the directories if '/' in post_filename: post_filename = post_filename.split('/')[-1] # filename of the post without any extension or path # This should also correspond to any index entry in # the posts cache post_url = remove_eol(post_filename) post_url = post_url.replace('.json', '').strip() # read the post from file full_post_filename = \ locate_post(base_dir, nickname, domain, post_url, False) if not full_post_filename: print('Unable to locate post for newswire ' + post_url) ctr += 1 if ctr >= max_blogs_per_account: break continue post_json_object = None if full_post_filename: post_json_object = load_json(full_post_filename) if _is_newswire_blog_post(post_json_object): published = post_json_object['object']['published'] published = published.replace('T', ' ') published = published.replace('Z', '+00:00') votes = [] if os.path.isfile(full_post_filename + '.votes'): votes = load_json(full_post_filename + '.votes') content = \ get_base_content_from_post(post_json_object, system_language) description = first_paragraph_from_string(content) description = remove_html(description) tags_from_post = \ _get_hashtags_from_post(post_json_object) summary = post_json_object['object']['summary'] url2 = post_json_object['object']['url'] url_str = get_url_from_post(url2) url3 = remove_html(url_str) fediverse_handle = '' extra_links = [] _add_newswire_dict_entry(base_dir, newswire, published, summary, url3, votes, full_post_filename, description, moderated, False, tags_from_post, max_tags, session, debug, None, system_language, fediverse_handle, extra_links) ctr += 1 if ctr >= max_blogs_per_account: break except OSError as exc: print('EX: _add_account_blogs_to_newswire unable to read ' + index_filename + ' ' + str(exc)) def _add_blogs_to_newswire(base_dir: str, domain: str, newswire: {}, max_blogs_per_account: int, max_tags: int, system_language: str, session, debug: bool) -> None: """Adds blogs from each user account into the newswire """ moderation_dict = {} # go through each account dir_str = data_dir(base_dir) for _, dirs, _ in os.walk(dir_str): for handle in dirs: if not is_account_dir(handle): continue nickname = handle.split('@')[0] # has this account been suspended? if is_suspended(base_dir, nickname): continue handle_dir = acct_handle_dir(base_dir, handle) if os.path.isfile(handle_dir + '/.nonewswire'): continue # is there a blogs timeline for this account? account_dir = os.path.join(dir_str, handle) blogs_index = account_dir + '/tlblogs.index' if os.path.isfile(blogs_index): domain = handle.split('@')[1] _add_account_blogs_to_newswire(base_dir, nickname, domain, newswire, max_blogs_per_account, blogs_index, max_tags, system_language, session, debug) break # sort the moderation dict into chronological order, latest first sorted_moderation_dict = \ OrderedDict(sorted(moderation_dict.items(), reverse=True)) # save the moderation queue details for later display newswire_moderation_filename = \ data_dir(base_dir) + '/newswiremoderation.txt' if sorted_moderation_dict: save_json(sorted_moderation_dict, newswire_moderation_filename) else: # remove the file if there is nothing to moderate if os.path.isfile(newswire_moderation_filename): try: os.remove(newswire_moderation_filename) except OSError: print('EX: _add_blogs_to_newswire unable to delete ' + str(newswire_moderation_filename)) def get_dict_from_newswire(session, base_dir: str, domain: str, max_posts_per_source: int, max_feed_size_kb: int, max_tags: int, max_feed_item_size_kb: int, max_newswire_posts: int, max_categories_feed_item_size_kb: int, system_language: str, debug: bool, preferred_podcast_formats: [], timeout_sec: int) -> {}: """Gets rss feeds as a dictionary from newswire file """ subscriptions_filename = data_dir(base_dir) + '/newswire.txt' if not os.path.isfile(subscriptions_filename): return {} max_posts_per_source = 5 # add rss feeds rss_feed = [] try: with open(subscriptions_filename, 'r', encoding='utf-8') as fp_sub: rss_feed = fp_sub.readlines() except OSError: print('EX: get_dict_from_newswire unable to read ' + subscriptions_filename) result = {} for url in rss_feed: url = url.strip() # Does this contain a url? if '://' not in url: continue # is this a comment? if url.startswith('#'): continue # should this feed be moderated? moderated = False if '*' in url: moderated = True url = url.replace('*', '').strip() # should this feed content be mirrored? mirrored = False if '!' in url: mirrored = True url = url.replace('!', '').strip() items_list = get_rss(base_dir, domain, session, url, moderated, mirrored, max_posts_per_source, max_feed_size_kb, max_feed_item_size_kb, max_categories_feed_item_size_kb, debug, preferred_podcast_formats, timeout_sec, system_language) if items_list: for date_str, item in items_list.items(): result[date_str] = item time.sleep(4) # add blogs from each user account _add_blogs_to_newswire(base_dir, domain, result, max_posts_per_source, max_tags, system_language, session, debug) # sort into chronological order, latest first sorted_result = OrderedDict(sorted(result.items(), reverse=True)) # are there too many posts? If so then remove the oldest ones no_of_posts = len(sorted_result.items()) if no_of_posts > max_newswire_posts: ctr = 0 removals = [] for date_str, item in sorted_result.items(): ctr += 1 if ctr > max_newswire_posts: removals.append(date_str) for remov in removals: sorted_result.pop(remov) return sorted_result