Browse Source

fix #3, fix #1, add exception handling for weird edgecases

master
rra 3 years ago
parent
commit
162ecce865
  1. 129
      rss_aggregator.py

129
rss_aggregator.py

@ -23,10 +23,10 @@ def write_etag(feed_name, feed_data):
etag = '' etag = ''
modified = '' modified = ''
if 'etag' in data: if 'etag' in feed_data:
etag = data.etag etag = feed_data.etag
if 'modified' in data: if 'modified' in feed_data:
modified = data.modified modified = feed_data.modified
if etag or modified: if etag or modified:
with open(os.path.join('etags',feed_name +'.txt'),'w') as f: with open(os.path.join('etags',feed_name +'.txt'),'w') as f:
@ -56,11 +56,16 @@ def create_frontmatter(entry):
published = arrow.get(published) published = arrow.get(published)
if 'author' in entry:
author = entry.author
else:
author = ''
frontmatter = { frontmatter = {
'title':entry.title, 'title':entry.title,
'date': published.format(), 'date': published.format(),
'summary': '', 'summary': '',
'author': entry.author, 'author': author,
'original_link': entry.link 'original_link': entry.link
} }
@ -75,7 +80,10 @@ def create_post(post_dir, entry):
if not os.path.exists(post_dir): if not os.path.exists(post_dir):
os.makedirs(post_dir) os.makedirs(post_dir)
post_content = entry.content[0].value if 'content' in entry:
post_content = entry.content[0].value
else:
post_content = entry.summary
parsed_content = parse_posts(post_dir, post_content) parsed_content = parse_posts(post_dir, post_content)
@ -91,20 +99,16 @@ def grab_media(post_directory, url):
""" """
image = urlparse(url).path.split('/')[-1] image = urlparse(url).path.split('/')[-1]
try:
#TODO: stream is true is a conditional so we could check the headers for things, mimetype etc
response = requests.get(url, stream=True)
except Exception as e:
print(e)
return url
try: try:
if not os.path.exists(os.path.join(post_directory, image)): if not os.path.exists(os.path.join(post_directory, image)):
with open(os.path.join(post_directory, image), 'wb') as img_file: #TODO: stream is true is a conditional so we could check the headers for things, mimetype etc
shutil.copyfileobj(response.raw, img_file) response = requests.get(url, stream=True)
print('Downloaded cover image', image) if response.ok:
return image with open(os.path.join(post_directory, image), 'wb') as img_file:
return image shutil.copyfileobj(response.raw, img_file)
print('Downloaded cover image', image)
return image
return image
except Exception as e: except Exception as e:
print('Failed to download cover image', url) print('Failed to download cover image', url)
@ -120,14 +124,41 @@ def parse_posts(post_direntry, post_content):
soup = BeautifulSoup(post_content, "html.parser") soup = BeautifulSoup(post_content, "html.parser")
media = [] media = []
for img in soup(['img','object']): for img in soup(['img','object']):
local_image = grab_media(post_dir, img['src']) local_image = grab_media(post_dir, img['src'])
if img['src'] != local_image: if img['src'] != local_image:
print(img['src'], '->', local_image) print(img['src'], '->', local_image)
img['src'] = local_image img['src'] = local_image
return soup.decode() return soup.decode()
def grab_feed(feed_url):
"""
check whether feed has been updated
download & return it if it has
"""
feed_name = urlparse(feed_url).netloc
etag, modified = get_etag(feed_name)
try:
if modified:
data = feedparser.parse(feed_url, modified=modified)
elif etag:
data = feedparser.parse(feed_url, etag=etag)
else:
data = feedparser.parse(feed_url)
except Exception as e:
print('Error grabbing feed')
print(feed_name)
print(e)
return False
print(data.status, feed_url)
if data.status == 200:
#304 means the feed has not been modified since we last checked
write_etag(feed_name, data)
return data
return False
feed_urls = open('feeds_list.txt','r').read().splitlines() feed_urls = open('feeds_list.txt','r').read().splitlines()
@ -141,8 +172,8 @@ env = jinja2.Environment(
loader=jinja2.FileSystemLoader(os.path.curdir) loader=jinja2.FileSystemLoader(os.path.curdir)
) )
output_dir = os.environ.get('OUTPUT_DIR', '/home/r/Programming/lumbung.space/lumbung.space-web/content/posts/') #output_dir = os.environ.get('OUTPUT_DIR', '/home/r/Programming/lumbung.space/lumbung.space-web/content/posts/')
#output_dir = os.environ.get('OUTPUT_DIR', 'network/') output_dir = os.environ.get('OUTPUT_DIR', 'network/')
if not os.path.exists(output_dir): if not os.path.exists(output_dir):
os.makedirs(output_dir) os.makedirs(output_dir)
@ -150,42 +181,40 @@ if not os.path.exists(output_dir):
template = env.get_template('post_template.md') template = env.get_template('post_template.md')
for feed_url in feed_urls[7:]: for feed_url in feed_urls:
feed_name = urlparse(feed_url).netloc feed_name = urlparse(feed_url).netloc
etag, modified = get_etag(feed_name) feed_dir = os.path.join(output_dir, feed_name)
if modified:
data = feedparser.parse(feed_url, modified=modified)
elif etag:
data = feedparser.parse(feed_url, etag=etag)
else:
data = feedparser.parse(feed_url)
print(data.status, feed_url) if not os.path.exists(feed_dir):
os.makedirs(feed_dir)
if data.status == 200: existing_posts = os.listdir(feed_dir)
#write_etag(feed_url, data) data = grab_feed(feed_url)
# if 'title' in data.feed: if data:
# print('#'*10) for entry in data.entries:
# print(data.feed.title) # if 'tags' in entry:
# print('#'*10) # print(entry.title, entry.tags)
# print('\n')
# print('FEED KEYS') post_name = slugify(entry.title)
# print(data.keys()) post_dir = os.path.join(output_dir, feed_name, post_name)
# print('\n')
for entry in data.entries: if post_name not in existing_posts:
# print(entry.title) #if there is a blog entry we dont already have, make it
# print(entry.keys()) create_post(post_dir, entry)
# print('\n')
# # if 'tags' in entry: elif post_name in existing_posts:
# # print(entry.title, entry.tags) #if we already have it, update it
create_post(post_dir, entry)
post_dir = os.path.join(output_dir, feed_name, slugify(entry.title)) existing_posts.remove(post_name) # create list of posts which have not been returned by the feed
create_post(post_dir, entry)
for post in existing_posts:
#remove blog posts no longer returned by the RSS feed
print('deleted', post)
shutil.rmtree(os.path.join(feed_dir, slugify(post)))

Loading…
Cancel
Save