2020-05-07 20:16:55 +02:00
|
|
|
#!/usr/bin/python3
|
|
|
|
# -*- coding: utf-8 -*-
|
2020-05-23 01:45:40 +02:00
|
|
|
|
2020-05-07 20:16:55 +02:00
|
|
|
"""
|
2020-11-12 21:19:51 +01:00
|
|
|
RSS module for InkyCal Project
|
2020-05-07 20:16:55 +02:00
|
|
|
Copyright by aceisace
|
|
|
|
"""
|
2020-12-02 01:00:28 +01:00
|
|
|
import re
|
2020-05-07 20:16:55 +02:00
|
|
|
|
2020-05-23 01:45:40 +02:00
|
|
|
from inkycal.modules.template import inkycal_module
|
2020-05-12 19:35:08 +02:00
|
|
|
from inkycal.custom import *
|
2020-05-07 20:16:55 +02:00
|
|
|
|
2020-05-23 01:45:40 +02:00
|
|
|
from random import shuffle
|
2020-05-07 20:16:55 +02:00
|
|
|
try:
|
|
|
|
import feedparser
|
|
|
|
except ImportError:
|
|
|
|
print('feedparser is not installed! Please install with:')
|
2020-05-13 18:36:54 +02:00
|
|
|
print('pip3 install feedparser')
|
2020-05-07 20:16:55 +02:00
|
|
|
|
2020-05-23 01:45:40 +02:00
|
|
|
filename = os.path.basename(__file__).split('.py')[0]
|
|
|
|
logger = logging.getLogger(filename)
|
2020-05-07 20:16:55 +02:00
|
|
|
|
2020-11-10 22:48:04 +01:00
|
|
|
class Feeds(inkycal_module):
|
2020-05-12 19:35:08 +02:00
|
|
|
"""RSS class
|
2020-05-26 19:10:20 +02:00
|
|
|
parses rss/atom feeds from given urls
|
2020-05-12 19:35:08 +02:00
|
|
|
"""
|
|
|
|
|
2020-11-21 16:22:15 +01:00
|
|
|
name = "RSS / Atom - Display feeds from given RSS/ATOM feeds"
|
2020-11-09 17:51:15 +01:00
|
|
|
|
|
|
|
requires = {
|
2020-11-10 11:53:48 +01:00
|
|
|
"feed_urls" : {
|
2020-11-09 17:51:15 +01:00
|
|
|
"label":"Please enter ATOM or RSS feed URL/s, separated by a comma",
|
|
|
|
},
|
|
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
optional = {
|
2020-11-10 22:48:04 +01:00
|
|
|
|
2020-11-09 17:51:15 +01:00
|
|
|
"shuffle_feeds": {
|
|
|
|
"label": "Should the parsed RSS feeds be shuffled? (default=True)",
|
|
|
|
"options": [True, False],
|
|
|
|
"default": True
|
|
|
|
},
|
|
|
|
|
|
|
|
}
|
|
|
|
|
2020-11-10 11:53:48 +01:00
|
|
|
def __init__(self, config):
|
|
|
|
"""Initialize inkycal_feeds module"""
|
2020-05-26 19:10:20 +02:00
|
|
|
|
2020-11-10 11:53:48 +01:00
|
|
|
super().__init__(config)
|
2020-05-18 03:46:49 +02:00
|
|
|
|
2020-11-10 11:53:48 +01:00
|
|
|
config = config['config']
|
|
|
|
|
|
|
|
# Check if all required parameters are present
|
2020-11-09 17:51:15 +01:00
|
|
|
for param in self.requires:
|
2020-11-10 11:53:48 +01:00
|
|
|
if not param in config:
|
2020-11-29 14:51:19 +01:00
|
|
|
raise Exception(f'config is missing {param}')
|
2020-05-21 01:00:37 +02:00
|
|
|
|
2020-11-10 11:53:48 +01:00
|
|
|
# required parameters
|
2020-11-21 16:22:15 +01:00
|
|
|
if config["feed_urls"] and isinstance(config['feed_urls'], str):
|
|
|
|
self.feed_urls = config["feed_urls"].split(",")
|
|
|
|
else:
|
|
|
|
self.feed_urls = config["feed_urls"]
|
2020-05-23 01:45:40 +02:00
|
|
|
|
2020-11-10 11:53:48 +01:00
|
|
|
# optional parameters
|
2020-11-12 21:19:51 +01:00
|
|
|
self.shuffle_feeds = config["shuffle_feeds"]
|
2020-11-10 22:48:04 +01:00
|
|
|
|
2020-05-23 01:45:40 +02:00
|
|
|
# give an OK message
|
2020-11-29 14:51:19 +01:00
|
|
|
print(f'{filename} loaded')
|
2020-05-07 20:16:55 +02:00
|
|
|
|
2020-05-23 01:45:40 +02:00
|
|
|
def _validate(self):
|
|
|
|
"""Validate module-specific parameters"""
|
2020-11-09 17:51:15 +01:00
|
|
|
|
2020-05-23 01:45:40 +02:00
|
|
|
if not isinstance(self.shuffle_feeds, bool):
|
|
|
|
print('shuffle_feeds has to be a boolean: True/False')
|
|
|
|
|
2020-05-07 20:16:55 +02:00
|
|
|
|
|
|
|
def generate_image(self):
|
|
|
|
"""Generate image for this module"""
|
|
|
|
|
|
|
|
# Define new image size with respect to padding
|
2020-11-10 11:53:48 +01:00
|
|
|
im_width = int(self.width - (2 * self.padding_left))
|
|
|
|
im_height = int(self.height - (2 * self.padding_top))
|
2020-05-07 20:16:55 +02:00
|
|
|
im_size = im_width, im_height
|
2020-11-29 14:51:19 +01:00
|
|
|
logger.info(f'Image size: {im_size}')
|
2020-05-07 20:16:55 +02:00
|
|
|
|
|
|
|
# Create an image for black pixels and one for coloured pixels
|
2020-05-21 01:00:37 +02:00
|
|
|
im_black = Image.new('RGB', size = im_size, color = 'white')
|
2020-05-07 20:16:55 +02:00
|
|
|
im_colour = Image.new('RGB', size = im_size, color = 'white')
|
|
|
|
|
2020-05-12 19:35:08 +02:00
|
|
|
# Check if internet is available
|
|
|
|
if internet_available() == True:
|
2020-05-23 01:45:40 +02:00
|
|
|
logger.info('Connection test passed')
|
2020-05-12 19:35:08 +02:00
|
|
|
else:
|
2020-05-21 01:00:37 +02:00
|
|
|
raise Exception('Network could not be reached :/')
|
2020-05-12 19:35:08 +02:00
|
|
|
|
2020-11-10 11:53:48 +01:00
|
|
|
# Set some parameters for formatting feeds
|
2020-05-07 20:16:55 +02:00
|
|
|
line_spacing = 1
|
|
|
|
line_height = self.font.getsize('hg')[1] + line_spacing
|
|
|
|
line_width = im_width
|
|
|
|
max_lines = (im_height // (self.font.getsize('hg')[1] + line_spacing))
|
|
|
|
|
|
|
|
# Calculate padding from top so the lines look centralised
|
|
|
|
spacing_top = int( im_height % line_height / 2 )
|
|
|
|
|
|
|
|
# Calculate line_positions
|
|
|
|
line_positions = [
|
|
|
|
(0, spacing_top + _ * line_height ) for _ in range(max_lines)]
|
|
|
|
|
2020-11-10 11:53:48 +01:00
|
|
|
# Create list containing all feeds from all urls
|
2020-05-21 01:00:37 +02:00
|
|
|
parsed_feeds = []
|
2020-11-10 11:53:48 +01:00
|
|
|
for feeds in self.feed_urls:
|
2020-05-21 01:00:37 +02:00
|
|
|
text = feedparser.parse(feeds)
|
|
|
|
for posts in text.entries:
|
2020-12-02 01:00:28 +01:00
|
|
|
summary = posts.summary
|
|
|
|
parsed_feeds.append(
|
|
|
|
f"•{posts.title}: {re.sub('<[^<]+?>', '', posts.summary)}")
|
2020-05-07 20:16:55 +02:00
|
|
|
|
2020-05-21 01:00:37 +02:00
|
|
|
self._parsed_feeds = parsed_feeds
|
2020-05-07 20:16:55 +02:00
|
|
|
|
2020-05-21 01:00:37 +02:00
|
|
|
# Shuffle the list to prevent showing the same content
|
|
|
|
if self.shuffle_feeds == True:
|
|
|
|
shuffle(parsed_feeds)
|
2020-05-07 20:16:55 +02:00
|
|
|
|
2020-05-21 01:00:37 +02:00
|
|
|
# Trim down the list to the max number of lines
|
|
|
|
del parsed_feeds[max_lines:]
|
|
|
|
|
|
|
|
# Wrap long text from feeds (line-breaking)
|
|
|
|
flatten = lambda z: [x for y in z for x in y]
|
|
|
|
filtered_feeds, counter = [], 0
|
|
|
|
|
|
|
|
for posts in parsed_feeds:
|
|
|
|
wrapped = text_wrap(posts, font = self.font, max_width = line_width)
|
|
|
|
counter += len(wrapped)
|
|
|
|
if counter < max_lines:
|
|
|
|
filtered_feeds.append(wrapped)
|
|
|
|
filtered_feeds = flatten(filtered_feeds)
|
|
|
|
self._filtered_feeds = filtered_feeds
|
|
|
|
|
2020-11-10 11:53:48 +01:00
|
|
|
logger.debug(f'filtered feeds -> {filtered_feeds}')
|
|
|
|
|
2020-05-21 01:00:37 +02:00
|
|
|
# Check if feeds could be parsed and can be displayed
|
|
|
|
if len(filtered_feeds) == 0 and len(parsed_feeds) > 0:
|
|
|
|
print('Feeds could be parsed, but the text is too long to be displayed:/')
|
|
|
|
elif len(filtered_feeds) == 0 and len(parsed_feeds) == 0:
|
|
|
|
print('No feeds could be parsed :/')
|
|
|
|
else:
|
2020-11-10 11:53:48 +01:00
|
|
|
# Write feeds on image
|
2020-05-07 20:16:55 +02:00
|
|
|
for _ in range(len(filtered_feeds)):
|
|
|
|
write(im_black, line_positions[_], (line_width, line_height),
|
|
|
|
filtered_feeds[_], font = self.font, alignment= 'left')
|
|
|
|
|
2020-11-13 10:14:30 +01:00
|
|
|
# return images
|
2020-11-09 17:51:15 +01:00
|
|
|
return im_black, im_colour
|
2020-05-07 20:16:55 +02:00
|
|
|
|
2020-05-23 01:45:40 +02:00
|
|
|
if __name__ == '__main__':
|
2020-11-29 14:51:19 +01:00
|
|
|
print(f'running {filename} in standalone/debug mode')
|