forked from tanner/qotnews
fix news.py feed.
This commit is contained in:
parent
2dbc702b40
commit
1658346aa9
|
@ -1,7 +1,7 @@
|
||||||
import logging
|
import logging
|
||||||
logging.basicConfig(
|
logging.basicConfig(
|
||||||
format='%(asctime)s - %(name)s - %(levelname)s - %(message)s',
|
format='%(asctime)s - %(name)s - %(levelname)s - %(message)s',
|
||||||
level=logging.DEBUG)
|
level=logging.ERROR)
|
||||||
|
|
||||||
if __name__ == '__main__':
|
if __name__ == '__main__':
|
||||||
import sys
|
import sys
|
||||||
|
@ -22,7 +22,7 @@ USER_AGENT = 'Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:77.0) Gecko/20100101
|
||||||
|
|
||||||
def unix(date_str, tz=None):
|
def unix(date_str, tz=None):
|
||||||
try:
|
try:
|
||||||
dt = dateutil.parser.parse(date_str, f)
|
dt = dateutil.parser.parse(date_str)
|
||||||
if tz:
|
if tz:
|
||||||
dt = pytz.timezone(tz).localize(dt)
|
dt = pytz.timezone(tz).localize(dt)
|
||||||
return int(dt.timestamp())
|
return int(dt.timestamp())
|
||||||
|
@ -158,6 +158,13 @@ class _Base:
|
||||||
return False
|
return False
|
||||||
return s
|
return s
|
||||||
|
|
||||||
|
def get_sitemap_date(a):
|
||||||
|
if a.find('lastmod'):
|
||||||
|
return a.find('lastmod').text
|
||||||
|
if a.find('news:publication_date'):
|
||||||
|
return a.find('news:publication_date').text
|
||||||
|
return ''
|
||||||
|
|
||||||
class Sitemap(_Base):
|
class Sitemap(_Base):
|
||||||
def __init__(self, url, tz=None):
|
def __init__(self, url, tz=None):
|
||||||
self.tz = tz
|
self.tz = tz
|
||||||
|
@ -167,13 +174,12 @@ class Sitemap(_Base):
|
||||||
markup = xml(lambda x: self.sitemap_url)
|
markup = xml(lambda x: self.sitemap_url)
|
||||||
if not markup: return []
|
if not markup: return []
|
||||||
soup = BeautifulSoup(markup, features='lxml')
|
soup = BeautifulSoup(markup, features='lxml')
|
||||||
articles = soup.find('urlset').findAll('url')
|
sitemap = soup.find('urlset').findAll('url')
|
||||||
news = list(filter(None, [a if a.find('news:news') else None for a in articles]))
|
|
||||||
news = list(filter(None, [a if a.find('news:publication_date') else None for a in news]))
|
links = list(filter(None, [a if a.find('loc') else None for a in sitemap]))
|
||||||
articles = list(filter(None, [a if a.find('lastmod') else None for a in articles]))
|
links = list(filter(None, [a if get_sitemap_date(a) else None for a in links]))
|
||||||
links = articles + news
|
links.sort(key=lambda a: unix(get_sitemap_date(a)), reverse=True)
|
||||||
links.sort(key=lambda a: unix(a.find('lastmod')), reverse=True)
|
links = [x.find('loc').text for x in links] or []
|
||||||
links = [x.find('loc').text for x in articles] or []
|
|
||||||
links = list(set(links))
|
links = list(set(links))
|
||||||
if excludes:
|
if excludes:
|
||||||
links = list(filter(None, [None if any(e in link for e in excludes) else link for link in links]))
|
links = list(filter(None, [None if any(e in link for e in excludes) else link for link in links]))
|
||||||
|
@ -204,15 +210,15 @@ class Category(_Base):
|
||||||
# scratchpad so I can quickly develop the parser
|
# scratchpad so I can quickly develop the parser
|
||||||
if __name__ == '__main__':
|
if __name__ == '__main__':
|
||||||
print("Sitemap: Stuff")
|
print("Sitemap: Stuff")
|
||||||
site = Sitemap("https://www.stuff.co.nz/sitemap.xml")
|
site = Sitemap("https://www.stuff.co.nz/sitemap/news/sitemap.xml")
|
||||||
posts = site.feed()
|
posts = site.feed()
|
||||||
print(posts[:1])
|
print(posts[:5])
|
||||||
print(site.story(posts[0]))
|
print(site.story(posts[0]))
|
||||||
|
|
||||||
print("Category: RadioNZ Te Ao Māori")
|
print("Category: RadioNZ Te Ao Māori")
|
||||||
site = Category("https://www.rnz.co.nz/news/te-manu-korihi/")
|
site = Category("https://www.rnz.co.nz/news/te-manu-korihi/")
|
||||||
posts = site.feed()
|
posts = site.feed()
|
||||||
print(posts[:1])
|
print(posts[:5])
|
||||||
print(site.story(posts[0]))
|
print(site.story(posts[0]))
|
||||||
|
|
||||||
print("Sitemap: Newsroom")
|
print("Sitemap: Newsroom")
|
||||||
|
|
Loading…
Reference in New Issue
Block a user