Compare commits
7 Commits
57de076fec
...
master
Author | SHA1 | Date | |
---|---|---|---|
9ec61ea5bc | |||
bdc7a6c10d | |||
4858516b01 | |||
f10e6063fc | |||
249a616531 | |||
ab92bd5441 | |||
6b16a768a7 |
@@ -9,6 +9,7 @@ from bs4 import BeautifulSoup
|
||||
|
||||
import settings
|
||||
from feeds import hackernews, reddit, tildes, manual, lobsters
|
||||
import utils
|
||||
|
||||
INVALID_DOMAINS = ['youtube.com', 'bloomberg.com', 'wsj.com', 'sec.gov']
|
||||
TWO_DAYS = 60*60*24*2
|
||||
@@ -68,6 +69,7 @@ def get_content_type(url):
|
||||
def update_story(story, is_manual=False):
|
||||
res = {}
|
||||
|
||||
try:
|
||||
if story['source'] == 'hackernews':
|
||||
res = hackernews.story(story['ref'])
|
||||
elif story['source'] == 'lobsters':
|
||||
@@ -78,6 +80,10 @@ def update_story(story, is_manual=False):
|
||||
res = tildes.story(story['ref'])
|
||||
elif story['source'] == 'manual':
|
||||
res = manual.story(story['ref'])
|
||||
except BaseException as e:
|
||||
utils.alert_tanner('Problem updating {} story, ref {}: {}'.format(story['source'], story['ref'], str(e)))
|
||||
logging.exception(e)
|
||||
return False
|
||||
|
||||
if res:
|
||||
story.update(res) # join dicts
|
||||
@@ -100,6 +106,12 @@ def update_story(story, is_manual=False):
|
||||
logging.info(story['url'])
|
||||
return False
|
||||
|
||||
if 'trump' in story['title'].lower() or 'musk' in story['title'].lower():
|
||||
logging.info('Trump / Musk story, skipping')
|
||||
logging.info(story['url'])
|
||||
return False
|
||||
|
||||
|
||||
logging.info('Getting article ' + story['url'])
|
||||
story['text'] = get_article(story['url'])
|
||||
if not story['text']: return False
|
||||
|
@@ -12,7 +12,8 @@ import requests
|
||||
from utils import clean
|
||||
|
||||
API_TOPSTORIES = lambda x: 'https://hacker-news.firebaseio.com/v0/topstories.json'
|
||||
API_ITEM = lambda x : 'https://hn.algolia.com/api/v1/items/{}'.format(x)
|
||||
ALG_API_ITEM = lambda x : 'https://hn.algolia.com/api/v1/items/{}'.format(x)
|
||||
BHN_API_ITEM = lambda x : 'https://api.hnpwa.com/v0/item/{}.json'.format(x)
|
||||
|
||||
SITE_LINK = lambda x : 'https://news.ycombinator.com/item?id={}'.format(x)
|
||||
SITE_AUTHOR_LINK = lambda x : 'https://news.ycombinator.com/user?id={}'.format(x)
|
||||
@@ -42,7 +43,7 @@ def api(route, ref=None):
|
||||
def feed():
|
||||
return [str(x) for x in api(API_TOPSTORIES) or []]
|
||||
|
||||
def comment(i):
|
||||
def alg_comment(i):
|
||||
if 'author' not in i:
|
||||
return False
|
||||
|
||||
@@ -51,19 +52,19 @@ def comment(i):
|
||||
c['score'] = i.get('points', 0)
|
||||
c['date'] = i.get('created_at_i', 0)
|
||||
c['text'] = clean(i.get('text', '') or '')
|
||||
c['comments'] = [comment(j) for j in i['children']]
|
||||
c['comments'] = [alg_comment(j) for j in i['children']]
|
||||
c['comments'] = list(filter(bool, c['comments']))
|
||||
return c
|
||||
|
||||
def comment_count(i):
|
||||
def alg_comment_count(i):
|
||||
alive = 1 if i['author'] else 0
|
||||
return sum([comment_count(c) for c in i['comments']]) + alive
|
||||
return sum([alg_comment_count(c) for c in i['comments']]) + alive
|
||||
|
||||
def story(ref):
|
||||
r = api(API_ITEM, ref)
|
||||
def alg_story(ref):
|
||||
r = api(ALG_API_ITEM, ref)
|
||||
if not r:
|
||||
logging.info('Bad Hackernews API response.')
|
||||
return False
|
||||
logging.info('Bad Algolia Hackernews API response.')
|
||||
return None
|
||||
|
||||
if 'deleted' in r:
|
||||
logging.info('Story was deleted.')
|
||||
@@ -80,17 +81,85 @@ def story(ref):
|
||||
s['title'] = r.get('title', '')
|
||||
s['link'] = SITE_LINK(ref)
|
||||
s['url'] = r.get('url', '')
|
||||
s['comments'] = [comment(i) for i in r['children']]
|
||||
s['comments'] = [alg_comment(i) for i in r['children']]
|
||||
s['comments'] = list(filter(bool, s['comments']))
|
||||
s['num_comments'] = comment_count(s) - 1
|
||||
s['num_comments'] = alg_comment_count(s) - 1
|
||||
|
||||
if 'text' in r and r['text']:
|
||||
s['text'] = clean(r['text'] or '')
|
||||
|
||||
return s
|
||||
|
||||
def bhn_comment(i):
|
||||
if 'user' not in i:
|
||||
return False
|
||||
|
||||
c = {}
|
||||
c['author'] = i.get('user', '')
|
||||
c['score'] = 0 # Not present?
|
||||
c['date'] = i.get('time', 0)
|
||||
c['text'] = clean(i.get('content', '') or '')
|
||||
c['comments'] = [bhn_comment(j) for j in i['comments']]
|
||||
c['comments'] = list(filter(bool, c['comments']))
|
||||
return c
|
||||
|
||||
def bhn_story(ref):
|
||||
r = api(BHN_API_ITEM, ref)
|
||||
if not r:
|
||||
logging.info('Bad BetterHN Hackernews API response.')
|
||||
return None
|
||||
|
||||
if 'deleted' in r: # TODO: verify
|
||||
logging.info('Story was deleted.')
|
||||
return False
|
||||
elif r.get('dead', False):
|
||||
logging.info('Story was deleted.')
|
||||
return False
|
||||
elif r.get('type', '') != 'link':
|
||||
logging.info('Type "{}" is not "link".'.format(r.get('type', '')))
|
||||
return False
|
||||
|
||||
s = {}
|
||||
s['author'] = r.get('user', '')
|
||||
s['author_link'] = SITE_AUTHOR_LINK(r.get('user', ''))
|
||||
s['score'] = r.get('points', 0)
|
||||
s['date'] = r.get('time', 0)
|
||||
s['title'] = r.get('title', '')
|
||||
s['link'] = SITE_LINK(ref)
|
||||
s['url'] = r.get('url', '')
|
||||
if s['url'].startswith('item'):
|
||||
s['url'] = SITE_LINK(ref)
|
||||
s['comments'] = [bhn_comment(i) for i in r['comments']]
|
||||
s['comments'] = list(filter(bool, s['comments']))
|
||||
s['num_comments'] = r.get('comments_count', 0)
|
||||
|
||||
if 'content' in r and r['content']:
|
||||
s['text'] = clean(r['content'] or '')
|
||||
|
||||
return s
|
||||
|
||||
def story(ref):
|
||||
s = alg_story(ref)
|
||||
if s is None:
|
||||
s = bhn_story(ref)
|
||||
if not s:
|
||||
return False
|
||||
|
||||
|
||||
if s['score'] < 25 and s['num_comments'] < 10:
|
||||
logging.info('Score ({}) or num comments ({}) below threshold.'.format(s['score'], s['num_comments']))
|
||||
return False
|
||||
|
||||
return s
|
||||
|
||||
# scratchpad so I can quickly develop the parser
|
||||
if __name__ == '__main__':
|
||||
print(feed())
|
||||
#print(story(20763961))
|
||||
#print(story(20802050))
|
||||
|
||||
#print(story(42899834)) # type "job"
|
||||
#print(story(42900076)) # Ask HN
|
||||
#print(story(42898201)) # Show HN
|
||||
#print(story(42899703)) # normal
|
||||
print(story(42902678)) # bad title?
|
||||
|
@@ -50,7 +50,7 @@ def unix(date_str):
|
||||
def make_comment(i):
|
||||
c = {}
|
||||
try:
|
||||
c['author'] = i['commenting_user']['username']
|
||||
c['author'] = i['commenting_user']
|
||||
except KeyError:
|
||||
c['author'] = ''
|
||||
c['score'] = i.get('score', 0)
|
||||
@@ -87,7 +87,7 @@ def story(ref):
|
||||
|
||||
s = {}
|
||||
try:
|
||||
s['author'] = r['submitter_user']['username']
|
||||
s['author'] = r['submitter_user']
|
||||
s['author_link'] = SITE_AUTHOR_LINK(s['author'])
|
||||
except KeyError:
|
||||
s['author'] = ''
|
||||
@@ -103,6 +103,10 @@ def story(ref):
|
||||
s['comments'] = iter_comments(r['comments'])
|
||||
s['num_comments'] = r['comment_count']
|
||||
|
||||
if s['score'] < 15 and s['num_comments'] < 10:
|
||||
logging.info('Score ({}) or num comments ({}) below threshold.'.format(s['score'], s['num_comments']))
|
||||
return False
|
||||
|
||||
if 'description' in r and r['description']:
|
||||
s['text'] = clean(r['description'] or '')
|
||||
|
||||
|
@@ -120,7 +120,7 @@ def story(ref):
|
||||
logging.info('Group ({}) not in whitelist.'.format(s['group']))
|
||||
return False
|
||||
|
||||
if s['score'] < 15 and s['num_comments'] < 15:
|
||||
if s['score'] < 15 and s['num_comments'] < 10:
|
||||
logging.info('Score ({}) or num comments ({}) below threshold.'.format(s['score'], s['num_comments']))
|
||||
return False
|
||||
|
||||
|
@@ -1,6 +1,8 @@
|
||||
import database
|
||||
import search
|
||||
import sys
|
||||
import settings
|
||||
import logging
|
||||
|
||||
import json
|
||||
import requests
|
||||
@@ -21,7 +23,7 @@ def database_del_story(sid):
|
||||
|
||||
def search_del_story(sid):
|
||||
try:
|
||||
r = requests.delete(search.MEILI_URL + 'indexes/qotnews/documents/'+sid, timeout=2)
|
||||
r = requests.delete(settings.MEILI_URL + 'indexes/qotnews/documents/'+sid, timeout=2)
|
||||
if r.status_code != 202:
|
||||
raise Exception('Bad response code ' + str(r.status_code))
|
||||
return r.json()
|
||||
|
@@ -8,6 +8,14 @@ import string
|
||||
|
||||
from bleach.sanitizer import Cleaner
|
||||
|
||||
def alert_tanner(message):
|
||||
try:
|
||||
logging.info('Alerting Tanner: ' + message)
|
||||
params = dict(qotnews=message)
|
||||
requests.get('https://tbot.tannercollin.com/message', params=params, timeout=4)
|
||||
except BaseException as e:
|
||||
logging.error('Problem alerting Tanner: ' + str(e))
|
||||
|
||||
def gen_rand_id():
|
||||
return ''.join(random.choice(string.ascii_uppercase) for _ in range(4))
|
||||
|
||||
|
Reference in New Issue
Block a user