epicyon/newswire.py

1210 lines
43 KiB
Python
Raw Normal View History

2020-10-04 09:51:12 +00:00
__filename__ = "newswire.py"
__author__ = "Bob Mottram"
__license__ = "AGPL3+"
2021-01-26 10:07:42 +00:00
__version__ = "1.2.0"
2020-10-04 09:51:12 +00:00
__maintainer__ = "Bob Mottram"
2021-09-10 16:14:50 +00:00
__email__ = "bob@libreserver.org"
2020-10-04 09:51:12 +00:00
__status__ = "Production"
2021-06-26 11:27:14 +00:00
__module_group__ = "Web Interface Columns"
2020-10-04 09:51:12 +00:00
import os
2021-02-12 11:28:00 +00:00
import json
2020-10-04 09:51:12 +00:00
import requests
from socket import error as SocketError
import errno
from datetime import datetime
2020-11-22 15:33:11 +00:00
from datetime import timedelta
2020-11-22 19:09:35 +00:00
from datetime import timezone
2020-10-04 09:51:12 +00:00
from collections import OrderedDict
from utils import validPostDate
from categories import setHashtagCategory
from utils import dangerousSVG
2021-12-17 12:01:54 +00:00
from utils import getFavFilenameFromUrl
from utils import getBaseContentFromPost
from utils import hasObjectDict
from utils import firstParagraphFromString
2020-10-25 10:42:38 +00:00
from utils import isPublicPost
2020-10-05 11:11:48 +00:00
from utils import locatePost
from utils import loadJson
from utils import saveJson
2020-10-06 08:58:44 +00:00
from utils import isSuspended
from utils import containsInvalidChars
from utils import removeHtml
2021-07-04 17:55:29 +00:00
from utils import isAccountDir
2021-07-13 21:59:53 +00:00
from utils import acctDir
2021-08-14 11:13:39 +00:00
from utils import localActorUrl
from blocking import isBlockedDomain
2020-10-17 14:23:35 +00:00
from blocking import isBlockedHashtag
2020-10-17 16:08:07 +00:00
from filters import isFiltered
2021-12-17 10:04:18 +00:00
from session import downloadImageAnyMimeType
2020-10-04 09:51:12 +00:00
2020-10-16 12:11:05 +00:00
def _removeCDATA(text: str) -> str:
"""Removes any CDATA from the given text
"""
if 'CDATA[' in text:
text = text.split('CDATA[')[1]
if ']' in text:
text = text.split(']')[0]
return text
2021-12-25 17:09:22 +00:00
def rss2Header(http_prefix: str,
2020-10-04 12:29:07 +00:00
nickname: str, domainFull: str,
title: str, translate: {}) -> str:
2020-10-06 09:22:23 +00:00
"""Header for an RSS 2.0 feed
"""
2021-07-04 09:50:09 +00:00
rssStr = \
"<?xml version=\"1.0\" encoding=\"UTF-8\" ?>" + \
"<rss version=\"2.0\">" + \
'<channel>'
2020-10-13 17:14:57 +00:00
2020-10-04 12:29:07 +00:00
if title.startswith('News'):
2021-07-04 09:50:09 +00:00
rssStr += \
' <title>Newswire</title>' + \
2021-12-25 17:09:22 +00:00
' <link>' + http_prefix + '://' + domainFull + \
2020-10-04 12:29:07 +00:00
'/newswire.xml' + '</link>'
2020-10-13 17:14:57 +00:00
elif title.startswith('Site'):
2021-07-04 09:50:09 +00:00
rssStr += \
' <title>' + domainFull + '</title>' + \
2021-12-25 17:09:22 +00:00
' <link>' + http_prefix + '://' + domainFull + \
2020-10-13 17:17:17 +00:00
'/blog/rss.xml' + '</link>'
2020-10-04 12:29:07 +00:00
else:
2021-07-04 09:50:09 +00:00
rssStr += \
' <title>' + translate[title] + '</title>' + \
2021-08-14 11:13:39 +00:00
' <link>' + \
2021-12-25 17:09:22 +00:00
localActorUrl(http_prefix, nickname, domainFull) + \
2021-08-14 11:13:39 +00:00
'/rss.xml' + '</link>'
2020-10-04 12:29:07 +00:00
return rssStr
def rss2Footer() -> str:
2020-10-06 09:22:23 +00:00
"""Footer for an RSS 2.0 feed
"""
2021-07-04 09:50:09 +00:00
rssStr = '</channel></rss>'
2020-10-04 12:29:07 +00:00
return rssStr
def getNewswireTags(text: str, maxTags: int) -> []:
2020-10-16 19:49:34 +00:00
"""Returns a list of hashtags found in the given text
"""
2020-10-16 20:46:34 +00:00
if '#' not in text:
return []
2020-10-16 19:49:34 +00:00
if ' ' not in text:
return []
textSimplified = \
text.replace(',', ' ').replace(';', ' ').replace('- ', ' ')
textSimplified = textSimplified.replace('. ', ' ').strip()
if textSimplified.endswith('.'):
textSimplified = textSimplified[:len(textSimplified)-1]
words = textSimplified.split(' ')
tags = []
for wrd in words:
if not wrd.startswith('#'):
continue
if len(wrd) <= 1:
continue
if wrd in tags:
continue
tags.append(wrd)
if len(tags) >= maxTags:
break
2020-10-16 19:49:34 +00:00
return tags
def limitWordLengths(text: str, maxWordLength: int) -> str:
"""Limits the maximum length of words so that the newswire
column cannot become too wide
"""
if ' ' not in text:
return text
words = text.split(' ')
result = ''
for wrd in words:
if len(wrd) > maxWordLength:
wrd = wrd[:maxWordLength]
if result:
result += ' '
result += wrd
return result
2021-12-16 21:22:44 +00:00
def getNewswireFaviconUrl(url: str) -> str:
2021-12-16 20:57:30 +00:00
"""Returns a favicon url from the given article link
"""
if '://' not in url:
return '/newswire_favicon.ico'
if url.startswith('http://'):
if not (url.endswith('.onion') or url.endswith('.i2p')):
return '/newswire_favicon.ico'
domain = url.split('://')[1]
if '/' not in domain:
return url + '/favicon.ico'
else:
domain = domain.split('/')[0]
return url.split('://')[0] + '://' + domain + '/favicon.ico'
2021-12-25 16:17:53 +00:00
def _downloadNewswireFeedFavicon(session, base_dir: str,
2021-12-16 20:57:30 +00:00
link: str, debug: bool) -> bool:
"""Downloads the favicon for the given feed link
"""
2021-12-16 22:01:12 +00:00
favUrl = getNewswireFaviconUrl(link)
2021-12-16 20:57:30 +00:00
if '://' not in link:
return False
timeoutSec = 10
2021-12-16 23:32:13 +00:00
imageData, mimeType = \
2021-12-17 10:04:18 +00:00
downloadImageAnyMimeType(session, favUrl, timeoutSec, debug)
2021-12-17 10:15:05 +00:00
if not imageData or not mimeType:
2021-12-16 23:59:53 +00:00
return False
2021-12-17 12:01:54 +00:00
# update the favicon url
extensionsToMime = {
'ico': 'x-icon',
'png': 'png',
'jpg': 'jpeg',
'gif': 'gif',
'avif': 'avif',
'svg': 'svg+xml',
'webp': 'webp'
}
for ext, mimeExt in extensionsToMime.items():
if 'image/' + mimeExt in mimeType:
2021-12-17 12:10:06 +00:00
favUrl = favUrl.replace('.ico', '.' + ext)
2021-12-17 12:01:54 +00:00
break
# create cached favicons directory if needed
2021-12-25 16:17:53 +00:00
if not os.path.isdir(base_dir + '/favicons'):
os.mkdir(base_dir + '/favicons')
2021-12-17 12:01:54 +00:00
# check svg for dubious scripts
if favUrl.endswith('.svg'):
2021-12-19 18:29:43 +00:00
imageDataStr = str(imageData)
if dangerousSVG(imageDataStr, False):
return False
2021-12-17 12:01:54 +00:00
# save to the cache
2021-12-25 16:17:53 +00:00
favFilename = getFavFilenameFromUrl(base_dir, favUrl)
2021-12-17 12:01:54 +00:00
if os.path.isfile(favFilename):
2021-12-16 21:14:24 +00:00
return True
2021-12-16 20:57:30 +00:00
try:
2021-12-17 12:01:54 +00:00
with open(favFilename, 'wb+') as fp:
2021-12-16 20:57:30 +00:00
fp.write(imageData)
except OSError:
2021-12-17 12:01:54 +00:00
print('EX: failed writing favicon ' + favFilename)
2021-12-16 20:57:30 +00:00
return False
2021-12-16 20:57:30 +00:00
return True
2021-12-25 16:17:53 +00:00
def _addNewswireDictEntry(base_dir: str, domain: str,
newswire: {}, dateStr: str,
title: str, link: str,
votesStatus: str, postFilename: str,
description: str, moderated: bool,
mirrored: bool,
tags: [],
2021-12-16 20:57:30 +00:00
maxTags: int, session, debug: bool) -> None:
"""Update the newswire dictionary
"""
# remove any markup
title = removeHtml(title)
description = removeHtml(description)
allText = title + ' ' + description
2020-10-25 10:17:12 +00:00
# check that none of the text is filtered against
2021-12-25 16:17:53 +00:00
if isFiltered(base_dir, None, None, allText):
2020-10-17 16:08:07 +00:00
return
2020-10-25 10:17:12 +00:00
title = limitWordLengths(title, 13)
2020-10-25 10:17:12 +00:00
if tags is None:
tags = []
# extract hashtags from the text of the feed post
postTags = getNewswireTags(allText, maxTags)
# combine the tags into a single list
2020-10-25 12:57:14 +00:00
for tag in tags:
if tag in postTags:
continue
if len(postTags) < maxTags:
postTags.append(tag)
2020-10-25 10:17:12 +00:00
# check that no tags are blocked
2020-10-25 12:57:14 +00:00
for tag in postTags:
2021-12-25 16:17:53 +00:00
if isBlockedHashtag(base_dir, tag):
2020-10-25 10:18:07 +00:00
return
2020-10-25 10:17:12 +00:00
2021-12-25 16:17:53 +00:00
_downloadNewswireFeedFavicon(session, base_dir, link, debug)
2021-12-16 20:57:30 +00:00
2020-10-25 10:17:12 +00:00
newswire[dateStr] = [
title,
link,
votesStatus,
postFilename,
description,
moderated,
2020-10-25 12:57:14 +00:00
postTags,
2020-10-25 10:17:12 +00:00
mirrored
]
2021-06-20 11:28:35 +00:00
def _validFeedDate(pubDate: str, debug: bool = False) -> bool:
# convert from YY-MM-DD HH:MM:SS+00:00 to
# YY-MM-DDTHH:MM:SSZ
postDate = pubDate.replace(' ', 'T').replace('+00:00', 'Z')
2021-03-14 19:53:22 +00:00
return validPostDate(postDate, 90, debug)
2020-11-22 19:01:18 +00:00
def parseFeedDate(pubDate: str) -> str:
"""Returns a UTC date string based on the given date string
2020-11-22 18:14:40 +00:00
This tries a number of formats to see which work
"""
formats = ("%a, %d %b %Y %H:%M:%S %z",
2021-10-17 14:17:42 +00:00
"%a, %d %b %Y %H:%M:%S Z",
2021-09-07 19:09:41 +00:00
"%a, %d %b %Y %H:%M:%S GMT",
2021-10-17 14:24:21 +00:00
"%a, %d %b %Y %H:%M:%S EST",
"%a, %d %b %Y %H:%M:%S PST",
"%a, %d %b %Y %H:%M:%S AST",
"%a, %d %b %Y %H:%M:%S CST",
"%a, %d %b %Y %H:%M:%S MST",
"%a, %d %b %Y %H:%M:%S AKST",
"%a, %d %b %Y %H:%M:%S HST",
"%a, %d %b %Y %H:%M:%S UT",
2020-11-22 18:14:40 +00:00
"%Y-%m-%dT%H:%M:%SZ",
"%Y-%m-%dT%H:%M:%S%z")
publishedDate = None
for dateFormat in formats:
if ',' in pubDate and ',' not in dateFormat:
continue
if ',' not in pubDate and ',' in dateFormat:
continue
if 'Z' in pubDate and 'Z' not in dateFormat:
continue
if 'Z' not in pubDate and 'Z' in dateFormat:
continue
if 'EST' not in pubDate and 'EST' in dateFormat:
continue
2021-09-07 19:09:41 +00:00
if 'GMT' not in pubDate and 'GMT' in dateFormat:
continue
2020-11-22 18:14:40 +00:00
if 'EST' in pubDate and 'EST' not in dateFormat:
continue
if 'UT' not in pubDate and 'UT' in dateFormat:
continue
if 'UT' in pubDate and 'UT' not in dateFormat:
continue
try:
2021-10-29 19:17:41 +00:00
publishedDate = datetime.strptime(pubDate, dateFormat)
2020-11-22 18:14:40 +00:00
except BaseException:
continue
if publishedDate:
if pubDate.endswith(' EST'):
hoursAdded = timedelta(hours=5)
publishedDate = publishedDate + hoursAdded
break
2020-11-22 19:01:18 +00:00
pubDateStr = None
if publishedDate:
2020-11-22 20:33:24 +00:00
offset = publishedDate.utcoffset()
2020-11-22 20:37:08 +00:00
if offset:
publishedDate = publishedDate - offset
2020-11-22 19:09:35 +00:00
# convert local date to UTC
publishedDate = publishedDate.replace(tzinfo=timezone.utc)
2020-11-22 19:01:18 +00:00
pubDateStr = str(publishedDate)
if not pubDateStr.endswith('+00:00'):
pubDateStr += '+00:00'
2021-09-07 19:33:27 +00:00
else:
print('WARN: unrecognized date format: ' + pubDate)
2020-11-22 19:01:18 +00:00
return pubDateStr
2020-11-22 18:14:40 +00:00
2021-12-25 16:17:53 +00:00
def loadHashtagCategories(base_dir: str, language: str) -> None:
"""Loads an rss file containing hashtag categories
"""
2021-12-25 16:17:53 +00:00
hashtagCategoriesFilename = base_dir + '/categories.xml'
if not os.path.isfile(hashtagCategoriesFilename):
hashtagCategoriesFilename = \
2021-12-25 16:17:53 +00:00
base_dir + '/defaultcategories/' + language + '.xml'
if not os.path.isfile(hashtagCategoriesFilename):
return
with open(hashtagCategoriesFilename, 'r') as fp:
xmlStr = fp.read()
2021-12-25 16:17:53 +00:00
_xml2StrToHashtagCategories(base_dir, xmlStr, 1024, True)
2021-12-25 16:17:53 +00:00
def _xml2StrToHashtagCategories(base_dir: str, xmlStr: str,
maxCategoriesFeedItemSizeKb: int,
2021-06-20 11:28:35 +00:00
force: bool = False) -> None:
2020-12-02 16:18:36 +00:00
"""Updates hashtag categories based upon an rss feed
"""
rssItems = xmlStr.split('<item>')
maxBytes = maxCategoriesFeedItemSizeKb * 1024
for rssItem in rssItems:
if not rssItem:
continue
if len(rssItem) > maxBytes:
print('WARN: rss categories feed item is too big')
continue
if '<title>' not in rssItem:
continue
if '</title>' not in rssItem:
continue
if '<description>' not in rssItem:
continue
if '</description>' not in rssItem:
continue
categoryStr = rssItem.split('<title>')[1]
categoryStr = categoryStr.split('</title>')[0].strip()
if not categoryStr:
continue
2020-12-03 10:12:09 +00:00
if 'CDATA' in categoryStr:
continue
2020-12-02 16:18:36 +00:00
hashtagListStr = rssItem.split('<description>')[1]
hashtagListStr = hashtagListStr.split('</description>')[0].strip()
if not hashtagListStr:
continue
2020-12-03 10:12:09 +00:00
if 'CDATA' in hashtagListStr:
continue
2020-12-02 16:18:36 +00:00
hashtagList = hashtagListStr.split(' ')
2021-12-25 16:17:53 +00:00
if not isBlockedHashtag(base_dir, categoryStr):
for hashtag in hashtagList:
2021-12-25 16:17:53 +00:00
setHashtagCategory(base_dir, hashtag, categoryStr,
False, force)
2020-12-02 16:18:36 +00:00
2021-12-25 16:17:53 +00:00
def _xml2StrToDict(base_dir: str, domain: str, xmlStr: str,
moderated: bool, mirrored: bool,
maxPostsPerSource: int,
2021-12-25 18:57:13 +00:00
max_feed_item_size_kb: int,
2021-12-16 20:57:30 +00:00
maxCategoriesFeedItemSizeKb: int,
session, debug: bool) -> {}:
2020-12-14 14:22:44 +00:00
"""Converts an xml RSS 2.0 string to a dictionary
2020-10-04 09:51:12 +00:00
"""
if '<item>' not in xmlStr:
return {}
result = {}
2020-12-09 10:38:09 +00:00
# is this an rss feed containing hashtag categories?
2020-12-02 16:18:36 +00:00
if '<title>#categories</title>' in xmlStr:
2021-12-25 16:17:53 +00:00
_xml2StrToHashtagCategories(base_dir, xmlStr,
maxCategoriesFeedItemSizeKb)
2020-12-02 16:18:36 +00:00
return {}
2020-12-09 10:38:09 +00:00
2020-10-04 09:51:12 +00:00
rssItems = xmlStr.split('<item>')
postCtr = 0
2021-12-25 18:57:13 +00:00
maxBytes = max_feed_item_size_kb * 1024
2020-10-04 09:51:12 +00:00
for rssItem in rssItems:
2020-11-27 22:43:34 +00:00
if not rssItem:
continue
if len(rssItem) > maxBytes:
print('WARN: rss feed item is too big')
continue
2020-10-04 09:51:12 +00:00
if '<title>' not in rssItem:
continue
if '</title>' not in rssItem:
continue
if '<link>' not in rssItem:
continue
if '</link>' not in rssItem:
continue
if '<pubDate>' not in rssItem:
continue
if '</pubDate>' not in rssItem:
continue
title = rssItem.split('<title>')[1]
title = _removeCDATA(title.split('</title>')[0])
2021-02-13 21:48:24 +00:00
title = removeHtml(title)
2020-10-07 12:05:49 +00:00
description = ''
if '<description>' in rssItem and '</description>' in rssItem:
description = rssItem.split('<description>')[1]
2021-01-11 21:54:25 +00:00
description = removeHtml(description.split('</description>')[0])
else:
if '<media:description>' in rssItem and \
'</media:description>' in rssItem:
description = rssItem.split('<media:description>')[1]
description = description.split('</media:description>')[0]
2021-01-11 21:54:25 +00:00
description = removeHtml(description)
2020-10-04 09:51:12 +00:00
link = rssItem.split('<link>')[1]
link = link.split('</link>')[0]
if '://' not in link:
continue
2020-10-17 20:53:36 +00:00
itemDomain = link.split('://')[1]
if '/' in itemDomain:
itemDomain = itemDomain.split('/')[0]
2021-12-25 16:17:53 +00:00
if isBlockedDomain(base_dir, itemDomain):
continue
2020-10-04 09:51:12 +00:00
pubDate = rssItem.split('<pubDate>')[1]
pubDate = pubDate.split('</pubDate>')[0]
2020-11-22 18:14:40 +00:00
2020-11-22 19:01:18 +00:00
pubDateStr = parseFeedDate(pubDate)
if pubDateStr:
if _validFeedDate(pubDateStr):
postFilename = ''
votesStatus = []
2021-12-25 16:17:53 +00:00
_addNewswireDictEntry(base_dir, domain,
result, pubDateStr,
title, link,
votesStatus, postFilename,
2021-02-12 11:28:00 +00:00
description, moderated,
2021-12-16 20:57:30 +00:00
mirrored, [], 32, session, debug)
postCtr += 1
if postCtr >= maxPostsPerSource:
break
2020-11-27 22:43:34 +00:00
if postCtr > 0:
2021-02-12 11:28:00 +00:00
print('Added ' + str(postCtr) +
' rss 2.0 feed items to newswire')
2020-10-04 09:51:12 +00:00
return result
2021-12-25 16:17:53 +00:00
def _xml1StrToDict(base_dir: str, domain: str, xmlStr: str,
moderated: bool, mirrored: bool,
maxPostsPerSource: int,
2021-12-25 18:57:13 +00:00
max_feed_item_size_kb: int,
2021-12-16 20:57:30 +00:00
maxCategoriesFeedItemSizeKb: int,
session, debug: bool) -> {}:
2020-12-14 14:22:44 +00:00
"""Converts an xml RSS 1.0 string to a dictionary
https://validator.w3.org/feed/docs/rss1.html
"""
2020-12-14 17:18:16 +00:00
itemStr = '<item'
if itemStr not in xmlStr:
2020-12-14 14:22:44 +00:00
return {}
result = {}
# is this an rss feed containing hashtag categories?
if '<title>#categories</title>' in xmlStr:
2021-12-25 16:17:53 +00:00
_xml2StrToHashtagCategories(base_dir, xmlStr,
maxCategoriesFeedItemSizeKb)
2020-12-14 14:22:44 +00:00
return {}
2020-12-14 17:18:16 +00:00
rssItems = xmlStr.split(itemStr)
2020-12-14 14:22:44 +00:00
postCtr = 0
2021-12-25 18:57:13 +00:00
maxBytes = max_feed_item_size_kb * 1024
2020-12-14 14:22:44 +00:00
for rssItem in rssItems:
if not rssItem:
continue
if len(rssItem) > maxBytes:
2020-12-14 17:18:16 +00:00
print('WARN: rss 1.0 feed item is too big')
continue
if rssItem.startswith('s>'):
2020-12-14 14:22:44 +00:00
continue
if '<title>' not in rssItem:
continue
if '</title>' not in rssItem:
continue
if '<link>' not in rssItem:
continue
if '</link>' not in rssItem:
continue
if '<dc:date>' not in rssItem:
continue
if '</dc:date>' not in rssItem:
continue
title = rssItem.split('<title>')[1]
title = _removeCDATA(title.split('</title>')[0])
2021-02-13 21:48:24 +00:00
title = removeHtml(title)
2020-12-14 14:22:44 +00:00
description = ''
if '<description>' in rssItem and '</description>' in rssItem:
description = rssItem.split('<description>')[1]
2021-01-11 21:54:25 +00:00
description = removeHtml(description.split('</description>')[0])
2020-12-14 14:22:44 +00:00
else:
if '<media:description>' in rssItem and \
'</media:description>' in rssItem:
description = rssItem.split('<media:description>')[1]
description = description.split('</media:description>')[0]
2021-01-11 21:54:25 +00:00
description = removeHtml(description)
2020-12-14 14:22:44 +00:00
link = rssItem.split('<link>')[1]
link = link.split('</link>')[0]
if '://' not in link:
continue
itemDomain = link.split('://')[1]
if '/' in itemDomain:
itemDomain = itemDomain.split('/')[0]
2021-12-25 16:17:53 +00:00
if isBlockedDomain(base_dir, itemDomain):
2020-12-14 14:22:44 +00:00
continue
pubDate = rssItem.split('<dc:date>')[1]
pubDate = pubDate.split('</dc:date>')[0]
pubDateStr = parseFeedDate(pubDate)
if pubDateStr:
if _validFeedDate(pubDateStr):
postFilename = ''
votesStatus = []
2021-12-25 16:17:53 +00:00
_addNewswireDictEntry(base_dir, domain,
result, pubDateStr,
title, link,
votesStatus, postFilename,
2021-02-12 11:28:00 +00:00
description, moderated,
2021-12-16 20:57:30 +00:00
mirrored, [], 32, session, debug)
postCtr += 1
if postCtr >= maxPostsPerSource:
break
2020-12-14 14:22:44 +00:00
if postCtr > 0:
2021-02-12 11:28:00 +00:00
print('Added ' + str(postCtr) +
' rss 1.0 feed items to newswire')
2020-12-14 14:22:44 +00:00
return result
2021-12-25 16:17:53 +00:00
def _atomFeedToDict(base_dir: str, domain: str, xmlStr: str,
moderated: bool, mirrored: bool,
maxPostsPerSource: int,
2021-12-25 18:57:13 +00:00
max_feed_item_size_kb: int,
2021-12-16 20:57:30 +00:00
session, debug: bool) -> {}:
2020-10-10 12:24:14 +00:00
"""Converts an atom feed string to a dictionary
"""
if '<entry>' not in xmlStr:
return {}
result = {}
2020-11-22 12:41:54 +00:00
atomItems = xmlStr.split('<entry>')
postCtr = 0
2021-12-25 18:57:13 +00:00
maxBytes = max_feed_item_size_kb * 1024
2020-11-22 12:41:54 +00:00
for atomItem in atomItems:
2020-11-27 22:43:34 +00:00
if not atomItem:
continue
2020-11-22 12:41:54 +00:00
if len(atomItem) > maxBytes:
print('WARN: atom feed item is too big')
continue
2020-11-22 12:41:54 +00:00
if '<title>' not in atomItem:
2020-10-10 12:24:14 +00:00
continue
2020-11-22 12:41:54 +00:00
if '</title>' not in atomItem:
2020-10-10 12:24:14 +00:00
continue
2020-11-22 12:41:54 +00:00
if '<link>' not in atomItem:
2020-10-10 12:24:14 +00:00
continue
2020-11-22 12:41:54 +00:00
if '</link>' not in atomItem:
2020-10-10 12:24:14 +00:00
continue
2020-11-22 12:41:54 +00:00
if '<updated>' not in atomItem:
2020-10-10 12:24:14 +00:00
continue
2020-11-22 12:41:54 +00:00
if '</updated>' not in atomItem:
2020-10-10 12:24:14 +00:00
continue
2020-11-22 12:41:54 +00:00
title = atomItem.split('<title>')[1]
title = _removeCDATA(title.split('</title>')[0])
2021-02-13 21:48:24 +00:00
title = removeHtml(title)
2020-10-10 12:24:14 +00:00
description = ''
2020-11-22 12:41:54 +00:00
if '<summary>' in atomItem and '</summary>' in atomItem:
description = atomItem.split('<summary>')[1]
2021-01-11 21:54:25 +00:00
description = removeHtml(description.split('</summary>')[0])
else:
2020-11-22 12:41:54 +00:00
if '<media:description>' in atomItem and \
'</media:description>' in atomItem:
description = atomItem.split('<media:description>')[1]
description = description.split('</media:description>')[0]
2021-01-11 21:54:25 +00:00
description = removeHtml(description)
2020-11-22 12:41:54 +00:00
link = atomItem.split('<link>')[1]
2020-10-10 12:24:14 +00:00
link = link.split('</link>')[0]
if '://' not in link:
continue
2020-10-17 20:53:36 +00:00
itemDomain = link.split('://')[1]
if '/' in itemDomain:
itemDomain = itemDomain.split('/')[0]
2021-12-25 16:17:53 +00:00
if isBlockedDomain(base_dir, itemDomain):
continue
2020-11-22 12:41:54 +00:00
pubDate = atomItem.split('<updated>')[1]
2020-10-10 12:24:14 +00:00
pubDate = pubDate.split('</updated>')[0]
2020-11-22 18:14:40 +00:00
2020-11-22 19:01:18 +00:00
pubDateStr = parseFeedDate(pubDate)
if pubDateStr:
if _validFeedDate(pubDateStr):
postFilename = ''
votesStatus = []
2021-12-25 16:17:53 +00:00
_addNewswireDictEntry(base_dir, domain,
result, pubDateStr,
title, link,
votesStatus, postFilename,
2021-02-12 11:28:00 +00:00
description, moderated,
2021-12-16 20:57:30 +00:00
mirrored, [], 32, session, debug)
2021-02-12 11:28:00 +00:00
postCtr += 1
if postCtr >= maxPostsPerSource:
break
if postCtr > 0:
print('Added ' + str(postCtr) +
' atom feed items to newswire')
return result
2021-12-25 16:17:53 +00:00
def _jsonFeedV1ToDict(base_dir: str, domain: str, xmlStr: str,
2021-02-12 11:30:23 +00:00
moderated: bool, mirrored: bool,
maxPostsPerSource: int,
2021-12-25 18:57:13 +00:00
max_feed_item_size_kb: int,
2021-12-16 20:57:30 +00:00
session, debug: bool) -> {}:
2021-02-12 11:28:00 +00:00
"""Converts a json feed string to a dictionary
2021-02-12 11:46:26 +00:00
See https://jsonfeed.org/version/1.1
2021-02-12 11:28:00 +00:00
"""
2021-02-12 11:46:26 +00:00
if '"items"' not in xmlStr:
2021-02-12 11:28:00 +00:00
return {}
try:
feedJson = json.loads(xmlStr)
except BaseException:
2021-10-29 18:48:15 +00:00
print('EX: _jsonFeedV1ToDict unable to load json ' + str(xmlStr))
2021-02-12 11:28:00 +00:00
return {}
2021-12-25 18:57:13 +00:00
maxBytes = max_feed_item_size_kb * 1024
2021-02-12 11:28:00 +00:00
if not feedJson.get('version'):
return {}
2021-02-12 11:46:26 +00:00
if not feedJson['version'].startswith('https://jsonfeed.org/version/1'):
2021-02-12 11:28:00 +00:00
return {}
if not feedJson.get('items'):
return {}
if not isinstance(feedJson['items'], list):
return {}
2021-02-12 11:50:05 +00:00
postCtr = 0
2021-02-12 11:47:49 +00:00
result = {}
2021-02-12 11:28:00 +00:00
for jsonFeedItem in feedJson['items']:
if not jsonFeedItem:
continue
if not isinstance(jsonFeedItem, dict):
continue
if not jsonFeedItem.get('url'):
continue
if not isinstance(jsonFeedItem['url'], str):
continue
if not jsonFeedItem.get('date_published'):
if not jsonFeedItem.get('date_modified'):
continue
if not jsonFeedItem.get('content_text'):
if not jsonFeedItem.get('content_html'):
continue
if jsonFeedItem.get('content_html'):
if not isinstance(jsonFeedItem['content_html'], str):
continue
title = removeHtml(jsonFeedItem['content_html'])
else:
if not isinstance(jsonFeedItem['content_text'], str):
continue
2021-02-12 12:02:09 +00:00
title = removeHtml(jsonFeedItem['content_text'])
2021-02-12 11:28:00 +00:00
if len(title) > maxBytes:
print('WARN: json feed title is too long')
continue
description = ''
if jsonFeedItem.get('description'):
if not isinstance(jsonFeedItem['description'], str):
continue
2021-02-12 12:02:09 +00:00
description = removeHtml(jsonFeedItem['description'])
2021-02-12 11:28:00 +00:00
if len(description) > maxBytes:
print('WARN: json feed description is too long')
continue
2021-02-12 12:09:16 +00:00
if jsonFeedItem.get('tags'):
2021-02-12 12:09:48 +00:00
if isinstance(jsonFeedItem['tags'], list):
2021-02-12 12:09:16 +00:00
for tagName in jsonFeedItem['tags']:
if not isinstance(tagName, str):
continue
if ' ' in tagName:
continue
if not tagName.startswith('#'):
tagName = '#' + tagName
if tagName not in description:
description += ' ' + tagName
2021-02-12 11:28:00 +00:00
link = jsonFeedItem['url']
if '://' not in link:
continue
if len(link) > maxBytes:
print('WARN: json feed link is too long')
continue
itemDomain = link.split('://')[1]
if '/' in itemDomain:
itemDomain = itemDomain.split('/')[0]
2021-12-25 16:17:53 +00:00
if isBlockedDomain(base_dir, itemDomain):
2021-02-12 11:28:00 +00:00
continue
if jsonFeedItem.get('date_published'):
if not isinstance(jsonFeedItem['date_published'], str):
continue
pubDate = jsonFeedItem['date_published']
else:
if not isinstance(jsonFeedItem['date_modified'], str):
continue
pubDate = jsonFeedItem['date_modified']
pubDateStr = parseFeedDate(pubDate)
if pubDateStr:
if _validFeedDate(pubDateStr):
postFilename = ''
votesStatus = []
2021-12-25 16:17:53 +00:00
_addNewswireDictEntry(base_dir, domain,
2021-02-12 11:28:00 +00:00
result, pubDateStr,
title, link,
votesStatus, postFilename,
description, moderated,
2021-12-16 20:57:30 +00:00
mirrored, [], 32, session, debug)
postCtr += 1
if postCtr >= maxPostsPerSource:
break
2020-11-27 22:43:34 +00:00
if postCtr > 0:
2021-02-12 11:28:00 +00:00
print('Added ' + str(postCtr) +
' json feed items to newswire')
2020-10-10 12:24:14 +00:00
return result
2021-12-25 16:17:53 +00:00
def _atomFeedYTToDict(base_dir: str, domain: str, xmlStr: str,
moderated: bool, mirrored: bool,
maxPostsPerSource: int,
2021-12-25 18:57:13 +00:00
max_feed_item_size_kb: int,
2021-12-16 20:57:30 +00:00
session, debug: bool) -> {}:
2020-11-22 10:34:42 +00:00
"""Converts an atom-style YouTube feed string to a dictionary
"""
if '<entry>' not in xmlStr:
return {}
2021-12-25 16:17:53 +00:00
if isBlockedDomain(base_dir, 'www.youtube.com'):
2020-11-22 10:34:42 +00:00
return {}
result = {}
2020-11-22 12:41:54 +00:00
atomItems = xmlStr.split('<entry>')
2020-11-22 10:34:42 +00:00
postCtr = 0
2021-12-25 18:57:13 +00:00
maxBytes = max_feed_item_size_kb * 1024
2020-11-22 12:41:54 +00:00
for atomItem in atomItems:
2020-11-27 22:43:34 +00:00
if not atomItem:
continue
if not atomItem.strip():
continue
2020-11-22 12:41:54 +00:00
if len(atomItem) > maxBytes:
2020-11-22 10:34:42 +00:00
print('WARN: atom feed item is too big')
continue
2020-11-22 12:41:54 +00:00
if '<title>' not in atomItem:
2020-11-22 10:34:42 +00:00
continue
2020-11-22 12:41:54 +00:00
if '</title>' not in atomItem:
2020-11-22 10:34:42 +00:00
continue
if '<published>' not in atomItem:
2020-11-22 10:34:42 +00:00
continue
if '</published>' not in atomItem:
2020-11-22 10:34:42 +00:00
continue
2020-11-22 12:41:54 +00:00
if '<yt:videoId>' not in atomItem:
2020-11-22 10:34:42 +00:00
continue
2020-11-22 12:41:54 +00:00
if '</yt:videoId>' not in atomItem:
2020-11-22 10:34:42 +00:00
continue
2020-11-22 12:41:54 +00:00
title = atomItem.split('<title>')[1]
title = _removeCDATA(title.split('</title>')[0])
2020-11-22 10:34:42 +00:00
description = ''
2020-11-22 12:41:54 +00:00
if '<media:description>' in atomItem and \
'</media:description>' in atomItem:
description = atomItem.split('<media:description>')[1]
2020-11-22 10:34:42 +00:00
description = description.split('</media:description>')[0]
2021-01-11 21:54:25 +00:00
description = removeHtml(description)
2020-11-22 12:41:54 +00:00
elif '<summary>' in atomItem and '</summary>' in atomItem:
description = atomItem.split('<summary>')[1]
2020-11-22 10:34:42 +00:00
description = description.split('</summary>')[0]
2021-01-11 21:54:25 +00:00
description = removeHtml(description)
2020-11-22 12:41:54 +00:00
link = atomItem.split('<yt:videoId>')[1]
2020-11-22 10:34:42 +00:00
link = link.split('</yt:videoId>')[0]
link = 'https://www.youtube.com/watch?v=' + link.strip()
pubDate = atomItem.split('<published>')[1]
pubDate = pubDate.split('</published>')[0]
2020-11-22 18:14:40 +00:00
2020-11-22 19:01:18 +00:00
pubDateStr = parseFeedDate(pubDate)
if pubDateStr:
if _validFeedDate(pubDateStr):
postFilename = ''
votesStatus = []
2021-12-25 16:17:53 +00:00
_addNewswireDictEntry(base_dir, domain,
result, pubDateStr,
title, link,
votesStatus, postFilename,
description, moderated, mirrored,
2021-12-16 20:57:30 +00:00
[], 32, session, debug)
postCtr += 1
if postCtr >= maxPostsPerSource:
break
2020-11-27 22:43:34 +00:00
if postCtr > 0:
print('Added ' + str(postCtr) + ' YouTube feed items to newswire')
2020-11-22 10:34:42 +00:00
return result
2021-12-25 16:17:53 +00:00
def _xmlStrToDict(base_dir: str, domain: str, xmlStr: str,
moderated: bool, mirrored: bool,
maxPostsPerSource: int,
2021-12-25 18:57:13 +00:00
max_feed_item_size_kb: int,
2021-12-16 20:57:30 +00:00
maxCategoriesFeedItemSizeKb: int,
session, debug: bool) -> {}:
2020-10-04 09:51:12 +00:00
"""Converts an xml string to a dictionary
"""
2020-11-22 16:10:58 +00:00
if '<yt:videoId>' in xmlStr and '<yt:channelId>' in xmlStr:
print('YouTube feed: reading')
2021-12-25 16:17:53 +00:00
return _atomFeedYTToDict(base_dir, domain,
xmlStr, moderated, mirrored,
2021-12-25 18:57:13 +00:00
maxPostsPerSource, max_feed_item_size_kb,
2021-12-16 20:57:30 +00:00
session, debug)
2020-11-22 16:10:58 +00:00
elif 'rss version="2.0"' in xmlStr:
2021-12-25 16:17:53 +00:00
return _xml2StrToDict(base_dir, domain,
xmlStr, moderated, mirrored,
2021-12-25 18:57:13 +00:00
maxPostsPerSource, max_feed_item_size_kb,
2021-12-16 20:57:30 +00:00
maxCategoriesFeedItemSizeKb,
session, debug)
2020-12-14 20:22:05 +00:00
elif '<?xml version="1.0"' in xmlStr:
2021-12-25 16:17:53 +00:00
return _xml1StrToDict(base_dir, domain,
xmlStr, moderated, mirrored,
2021-12-25 18:57:13 +00:00
maxPostsPerSource, max_feed_item_size_kb,
2021-12-16 20:57:30 +00:00
maxCategoriesFeedItemSizeKb,
session, debug)
elif 'xmlns="http://www.w3.org/2005/Atom"' in xmlStr:
2021-12-25 16:17:53 +00:00
return _atomFeedToDict(base_dir, domain,
xmlStr, moderated, mirrored,
2021-12-25 18:57:13 +00:00
maxPostsPerSource, max_feed_item_size_kb,
2021-12-16 20:57:30 +00:00
session, debug)
2021-02-12 11:28:00 +00:00
elif 'https://jsonfeed.org/version/1' in xmlStr:
2021-12-25 16:17:53 +00:00
return _jsonFeedV1ToDict(base_dir, domain,
2021-02-12 11:30:23 +00:00
xmlStr, moderated, mirrored,
2021-12-25 18:57:13 +00:00
maxPostsPerSource, max_feed_item_size_kb,
2021-12-16 20:57:30 +00:00
session, debug)
2020-10-04 09:51:12 +00:00
return {}
def _YTchannelToAtomFeed(url: str) -> str:
"""Converts a YouTube channel url into an atom feed url
"""
if 'youtube.com/channel/' not in url:
return url
2020-11-22 12:27:42 +00:00
channelId = url.split('youtube.com/channel/')[1].strip()
2020-11-22 12:36:21 +00:00
channelUrl = \
'https://www.youtube.com/feeds/videos.xml?channel_id=' + channelId
print('YouTube feed: ' + channelUrl)
return channelUrl
2021-12-25 16:17:53 +00:00
def getRSS(base_dir: str, domain: str, session, url: str,
moderated: bool, mirrored: bool,
maxPostsPerSource: int, maxFeedSizeKb: int,
2021-12-25 18:57:13 +00:00
max_feed_item_size_kb: int,
2021-12-16 20:57:30 +00:00
maxCategoriesFeedItemSizeKb: int, debug: bool) -> {}:
2020-10-04 09:51:12 +00:00
"""Returns an RSS url as a dict
"""
if not isinstance(url, str):
print('url: ' + str(url))
print('ERROR: getRSS url should be a string')
return None
headers = {
2020-12-14 20:22:05 +00:00
'Accept': 'text/xml, application/xml; charset=UTF-8'
2020-10-04 09:51:12 +00:00
}
params = None
sessionParams = {}
sessionHeaders = {}
if headers:
sessionHeaders = headers
if params:
sessionParams = params
sessionHeaders['User-Agent'] = \
'Mozilla/5.0 (X11; Linux x86_64; rv:81.0) Gecko/20100101 Firefox/81.0'
if not session:
print('WARN: no session specified for getRSS')
url = _YTchannelToAtomFeed(url)
2020-10-04 09:51:12 +00:00
try:
result = session.get(url, headers=sessionHeaders, params=sessionParams)
if result:
2020-10-16 12:03:56 +00:00
if int(len(result.text) / 1024) < maxFeedSizeKb and \
not containsInvalidChars(result.text):
2021-12-25 16:17:53 +00:00
return _xmlStrToDict(base_dir, domain, result.text,
moderated, mirrored,
maxPostsPerSource,
2021-12-25 18:57:13 +00:00
max_feed_item_size_kb,
2021-12-16 20:57:30 +00:00
maxCategoriesFeedItemSizeKb,
session, debug)
else:
2020-11-22 12:43:22 +00:00
print('WARN: feed is too large, ' +
'or contains invalid characters: ' + url)
2020-11-22 13:04:58 +00:00
else:
print('WARN: no result returned for feed ' + url)
2021-12-25 15:28:52 +00:00
except requests.exceptions.RequestException as ex:
2021-05-20 12:52:13 +00:00
print('WARN: getRSS failed\nurl: ' + str(url) + ', ' +
'headers: ' + str(sessionHeaders) + ', ' +
2021-12-25 15:28:52 +00:00
'params: ' + str(sessionParams) + ', ' + str(ex))
except ValueError as ex:
2021-05-20 12:52:13 +00:00
print('WARN: getRSS failed\nurl: ' + str(url) + ', ' +
'headers: ' + str(sessionHeaders) + ', ' +
2021-12-25 15:28:52 +00:00
'params: ' + str(sessionParams) + ', ' + str(ex))
except SocketError as ex:
if ex.errno == errno.ECONNRESET:
print('WARN: connection was reset during getRSS ' + str(ex))
2021-05-20 12:52:13 +00:00
else:
2021-12-25 15:28:52 +00:00
print('WARN: getRSS, ' + str(ex))
2020-10-04 09:51:12 +00:00
return None
2021-12-25 16:17:53 +00:00
def getRSSfromDict(base_dir: str, newswire: {},
2021-12-25 17:09:22 +00:00
http_prefix: str, domainFull: str,
2020-10-04 12:29:07 +00:00
title: str, translate: {}) -> str:
"""Returns an rss feed from the current newswire dict.
This allows other instances to subscribe to the same newswire
"""
2021-12-25 17:09:22 +00:00
rssStr = rss2Header(http_prefix,
2020-10-04 12:29:07 +00:00
None, domainFull,
'Newswire', translate)
2020-11-03 14:41:28 +00:00
if not newswire:
return ''
2020-10-04 12:29:07 +00:00
for published, fields in newswire.items():
2020-10-20 12:22:52 +00:00
if '+00:00' in published:
published = published.replace('+00:00', 'Z').strip()
published = published.replace(' ', 'T')
else:
publishedWithOffset = \
2020-10-20 12:37:32 +00:00
datetime.strptime(published, "%Y-%m-%d %H:%M:%S%z")
2020-10-20 12:22:52 +00:00
published = publishedWithOffset.strftime("%Y-%m-%dT%H:%M:%SZ")
2020-10-04 22:08:13 +00:00
try:
2020-10-04 22:12:27 +00:00
pubDate = datetime.strptime(published, "%Y-%m-%dT%H:%M:%SZ")
2021-12-25 15:28:52 +00:00
except Exception as ex:
print('WARN: Unable to convert date ' + published + ' ' + str(ex))
2020-10-04 22:08:13 +00:00
continue
2021-07-04 11:02:08 +00:00
rssStr += \
'<item>\n' + \
' <title>' + fields[0] + '</title>\n'
2021-01-11 21:54:25 +00:00
description = removeHtml(firstParagraphFromString(fields[4]))
rssStr += ' <description>' + description + '</description>\n'
2020-10-08 15:07:06 +00:00
url = fields[1]
2020-11-08 11:04:52 +00:00
if '://' not in url:
if domainFull not in url:
2021-12-25 17:09:22 +00:00
url = http_prefix + '://' + domainFull + url
2020-10-08 15:07:06 +00:00
rssStr += ' <link>' + url + '</link>\n'
2020-10-04 22:12:27 +00:00
2020-10-04 12:29:07 +00:00
rssDateStr = pubDate.strftime("%a, %d %b %Y %H:%M:%S UT")
2021-07-04 11:02:08 +00:00
rssStr += \
' <pubDate>' + rssDateStr + '</pubDate>\n' + \
'</item>\n'
2020-10-04 12:29:07 +00:00
rssStr += rss2Footer()
return rssStr
2021-12-25 22:09:19 +00:00
def _isNewswireBlogPost(post_json_object: {}) -> bool:
"""Is the given object a blog post?
2020-10-25 10:47:39 +00:00
There isn't any difference between a blog post and a newswire blog post
but we may here need to check for different properties than
isBlogPost does
"""
2021-12-25 22:09:19 +00:00
if not post_json_object:
return False
2021-12-25 22:09:19 +00:00
if not hasObjectDict(post_json_object):
return False
2021-12-25 22:09:19 +00:00
if post_json_object['object'].get('summary') and \
post_json_object['object'].get('url') and \
post_json_object['object'].get('content') and \
post_json_object['object'].get('published'):
return isPublicPost(post_json_object)
return False
2021-12-25 22:09:19 +00:00
def _getHashtagsFromPost(post_json_object: {}) -> []:
2020-10-16 20:13:23 +00:00
"""Returns a list of any hashtags within a post
"""
2021-12-25 22:09:19 +00:00
if not hasObjectDict(post_json_object):
2020-10-16 20:13:23 +00:00
return []
2021-12-25 22:09:19 +00:00
if not post_json_object['object'].get('tag'):
2020-10-16 20:13:23 +00:00
return []
2021-12-25 22:09:19 +00:00
if not isinstance(post_json_object['object']['tag'], list):
2020-10-16 20:13:23 +00:00
return []
tags = []
2021-12-25 22:09:19 +00:00
for tg in post_json_object['object']['tag']:
2020-10-16 20:13:23 +00:00
if not isinstance(tg, dict):
continue
if not tg.get('name'):
continue
if not tg.get('type'):
continue
if tg['type'] != 'Hashtag':
continue
if tg['name'] not in tags:
tags.append(tg['name'])
return tags
2021-12-25 16:17:53 +00:00
def _addAccountBlogsToNewswire(base_dir: str, nickname: str, domain: str,
newswire: {},
maxBlogsPerAccount: int,
indexFilename: str,
2021-12-25 23:03:28 +00:00
maxTags: int, system_language: str,
2021-12-16 20:57:30 +00:00
session, debug: bool) -> None:
2020-10-05 11:11:48 +00:00
"""Adds blogs for the given account to the newswire
"""
if not os.path.isfile(indexFilename):
return
# local blog entries are unmoderated by default
moderated = False
# local blogs can potentially be moderated
moderatedFilename = \
2021-12-25 16:17:53 +00:00
acctDir(base_dir, nickname, domain) + '/.newswiremoderated'
if os.path.isfile(moderatedFilename):
moderated = True
2020-10-05 11:11:48 +00:00
with open(indexFilename, 'r') as indexFile:
postFilename = 'start'
ctr = 0
while postFilename:
postFilename = indexFile.readline()
if postFilename:
# if this is a full path then remove the directories
if '/' in postFilename:
postFilename = postFilename.split('/')[-1]
# filename of the post without any extension or path
# This should also correspond to any index entry in
# the posts cache
postUrl = \
postFilename.replace('\n', '').replace('\r', '')
postUrl = postUrl.replace('.json', '').strip()
# read the post from file
fullPostFilename = \
2021-12-25 16:17:53 +00:00
locatePost(base_dir, nickname,
2020-10-05 11:11:48 +00:00
domain, postUrl, False)
2020-10-06 13:05:15 +00:00
if not fullPostFilename:
print('Unable to locate post for newswire ' + postUrl)
2020-10-06 13:05:15 +00:00
ctr += 1
if ctr >= maxBlogsPerAccount:
break
2020-10-06 13:34:04 +00:00
continue
2020-10-06 13:05:15 +00:00
2021-12-25 22:09:19 +00:00
post_json_object = None
2020-10-05 11:11:48 +00:00
if fullPostFilename:
2021-12-25 22:09:19 +00:00
post_json_object = loadJson(fullPostFilename)
if _isNewswireBlogPost(post_json_object):
published = post_json_object['object']['published']
published = published.replace('T', ' ')
published = published.replace('Z', '+00:00')
2020-10-06 20:17:34 +00:00
votes = []
if os.path.isfile(fullPostFilename + '.votes'):
votes = loadJson(fullPostFilename + '.votes')
content = \
2021-12-25 22:09:19 +00:00
getBaseContentFromPost(post_json_object,
2021-12-25 23:03:28 +00:00
system_language)
description = firstParagraphFromString(content)
2021-01-11 21:54:25 +00:00
description = removeHtml(description)
2021-12-25 22:09:19 +00:00
tagsFromPost = _getHashtagsFromPost(post_json_object)
summary = post_json_object['object']['summary']
2021-12-25 16:17:53 +00:00
_addNewswireDictEntry(base_dir, domain,
newswire, published,
2021-12-25 22:09:19 +00:00
summary,
post_json_object['object']['url'],
votes, fullPostFilename,
description, moderated, False,
tagsFromPost,
2021-12-16 20:57:30 +00:00
maxTags, session, debug)
2020-10-05 11:11:48 +00:00
ctr += 1
if ctr >= maxBlogsPerAccount:
break
2021-12-25 16:17:53 +00:00
def _addBlogsToNewswire(base_dir: str, domain: str, newswire: {},
maxBlogsPerAccount: int,
2021-12-25 23:03:28 +00:00
maxTags: int, system_language: str,
2021-12-16 20:57:30 +00:00
session, debug: bool) -> None:
"""Adds blogs from each user account into the newswire
2020-10-06 09:37:22 +00:00
"""
moderationDict = {}
2020-10-05 11:11:48 +00:00
# go through each account
2021-12-25 16:17:53 +00:00
for subdir, dirs, files in os.walk(base_dir + '/accounts'):
2020-10-05 11:11:48 +00:00
for handle in dirs:
2021-07-04 17:55:29 +00:00
if not isAccountDir(handle):
2020-10-05 11:11:48 +00:00
continue
2020-10-06 09:37:22 +00:00
nickname = handle.split('@')[0]
# has this account been suspended?
2021-12-25 16:17:53 +00:00
if isSuspended(base_dir, nickname):
2020-10-06 08:58:44 +00:00
continue
2021-12-25 16:17:53 +00:00
if os.path.isfile(base_dir + '/accounts/' + handle +
2020-10-06 21:28:40 +00:00
'/.nonewswire'):
continue
2020-10-05 11:11:48 +00:00
# is there a blogs timeline for this account?
2021-12-25 16:17:53 +00:00
accountDir = os.path.join(base_dir + '/accounts', handle)
2020-10-05 11:11:48 +00:00
blogsIndex = accountDir + '/tlblogs.index'
if os.path.isfile(blogsIndex):
domain = handle.split('@')[1]
2021-12-25 16:17:53 +00:00
_addAccountBlogsToNewswire(base_dir, nickname, domain,
newswire, maxBlogsPerAccount,
blogsIndex, maxTags,
2021-12-25 23:03:28 +00:00
system_language, session,
2021-12-16 20:57:30 +00:00
debug)
2020-12-13 22:13:45 +00:00
break
2020-10-05 11:11:48 +00:00
# sort the moderation dict into chronological order, latest first
sortedModerationDict = \
OrderedDict(sorted(moderationDict.items(), reverse=True))
2020-10-06 12:15:35 +00:00
# save the moderation queue details for later display
2021-12-25 16:17:53 +00:00
newswireModerationFilename = base_dir + '/accounts/newswiremoderation.txt'
if sortedModerationDict:
saveJson(sortedModerationDict, newswireModerationFilename)
else:
# remove the file if there is nothing to moderate
if os.path.isfile(newswireModerationFilename):
try:
os.remove(newswireModerationFilename)
2021-11-25 18:42:38 +00:00
except OSError:
2021-10-29 18:48:15 +00:00
print('EX: _addBlogsToNewswire unable to delete ' +
str(newswireModerationFilename))
2020-10-05 11:11:48 +00:00
2021-12-25 16:17:53 +00:00
def getDictFromNewswire(session, base_dir: str, domain: str,
maxPostsPerSource: int, maxFeedSizeKb: int,
2021-12-25 18:57:13 +00:00
maxTags: int, max_feed_item_size_kb: int,
2021-12-25 18:49:19 +00:00
max_newswire_posts: int,
maxCategoriesFeedItemSizeKb: int,
2021-12-25 23:03:28 +00:00
system_language: str, debug: bool) -> {}:
2020-10-04 09:59:55 +00:00
"""Gets rss feeds as a dictionary from newswire file
2020-10-04 09:51:12 +00:00
"""
2021-12-25 16:17:53 +00:00
subscriptionsFilename = base_dir + '/accounts/newswire.txt'
2020-10-04 09:51:12 +00:00
if not os.path.isfile(subscriptionsFilename):
return {}
maxPostsPerSource = 5
2020-10-05 11:11:48 +00:00
# add rss feeds
rssFeed = []
with open(subscriptionsFilename, 'r') as fp:
rssFeed = fp.readlines()
2020-10-04 09:51:12 +00:00
result = {}
for url in rssFeed:
url = url.strip()
# Does this contain a url?
2020-10-04 09:51:12 +00:00
if '://' not in url:
continue
# is this a comment?
2020-10-04 09:51:12 +00:00
if url.startswith('#'):
continue
# should this feed be moderated?
moderated = False
if '*' in url:
moderated = True
url = url.replace('*', '').strip()
# should this feed content be mirrored?
mirrored = False
if '!' in url:
mirrored = True
url = url.replace('!', '').strip()
2021-12-25 16:17:53 +00:00
itemsList = getRSS(base_dir, domain, session, url,
moderated, mirrored,
maxPostsPerSource, maxFeedSizeKb,
2021-12-25 18:57:13 +00:00
max_feed_item_size_kb,
2021-12-16 20:57:30 +00:00
maxCategoriesFeedItemSizeKb, debug)
2020-11-03 15:04:33 +00:00
if itemsList:
for dateStr, item in itemsList.items():
result[dateStr] = item
2020-10-05 11:11:48 +00:00
# add blogs from each user account
2021-12-25 16:17:53 +00:00
_addBlogsToNewswire(base_dir, domain, result,
2021-12-25 23:03:28 +00:00
maxPostsPerSource, maxTags, system_language,
2021-12-16 20:57:30 +00:00
session, debug)
2020-10-05 11:11:48 +00:00
# sort into chronological order, latest first
2020-10-04 21:45:46 +00:00
sortedResult = OrderedDict(sorted(result.items(), reverse=True))
# are there too many posts? If so then remove the oldest ones
noOfPosts = len(sortedResult.items())
2021-12-25 18:49:19 +00:00
if noOfPosts > max_newswire_posts:
2020-11-22 12:05:15 +00:00
ctr = 0
removals = []
for dateStr, item in sortedResult.items():
ctr += 1
2021-12-25 18:49:19 +00:00
if ctr > max_newswire_posts:
2020-11-22 12:05:15 +00:00
removals.append(dateStr)
for r in removals:
sortedResult.pop(r)
2020-10-04 09:51:12 +00:00
return sortedResult