Add requests timeouts and temporary logging

master
Tanner Collin 4 years ago
parent d614ad0743
commit 850b30e353
  1. 25
      apiserver/feed.py
  2. 8
      apiserver/feeds/tildes.py
  3. 2
      apiserver/server.py

@ -63,7 +63,7 @@ def get_first_image(text):
first_img = soup.find('img')
url = first_img['src']
headers = {'User-Agent': 'Twitterbot/1.0'}
length = requests.get(url, headers=headers).headers['content-length']
length = requests.get(url, headers=headers, timeout=4).headers['content-length']
if int(length) > 1000000: raise
return url
except:
@ -72,9 +72,15 @@ def get_first_image(text):
def get_content_type(url):
try:
headers = {'User-Agent': 'Twitterbot/1.0'}
return requests.get(url, headers=headers).headers['content-type']
return requests.get(url, headers=headers, timeout=2).headers['content-type']
except:
return ''
pass
try:
headers = {'User-Agent': 'Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:77.0) Gecko/20100101 Firefox/77.0'}
return requests.get(url, headers=headers, timeout=2).headers['content-type']
except:
return 'text/'
def update_story(story, is_manual=False):
res = {}
@ -90,26 +96,35 @@ def update_story(story, is_manual=False):
elif story['source'] == 'manual':
res = manual.story(story['ref'])
logging.info('Got story')
if res:
story.update(res) # join dicts
else:
logging.info('Article not ready yet')
logging.info('Story not ready yet')
return False
logging.info('story joined')
if story['date'] and not is_manual and story['date'] + TWO_DAYS < time.time():
logging.info('Article too old, removing')
logging.info('Story too old, removing')
return False
logging.info('story age good')
if story.get('url', '') and not story.get('text', ''):
logging.info('inside if')
if not get_content_type(story['url']).startswith('text/'):
logging.info('URL invalid file type / content type:')
logging.info(story['url'])
return False
logging.info('content type good')
if any([domain in story['url'] for domain in INVALID_DOMAINS]):
logging.info('URL invalid domain:')
logging.info(story['url'])
return False
logging.info('domain good')
logging.info('Getting article ' + story['url'])
story['text'] = get_article(story['url'])

@ -73,10 +73,14 @@ def story(ref):
html = api(API_ITEM(ref))
if not html: return False
logging.info('Got Tildes html')
soup = BeautifulSoup(html, features='html.parser')
a = soup.find('article', class_='topic-full')
if a is None: return False
logging.info('Got article tag')
h = a.find('header')
lu = h.find('a', class_='link-user')
@ -102,6 +106,8 @@ def story(ref):
ch = a.find('header', class_='topic-comments-header')
s['num_comments'] = int(ch.h2.string.split(' ')[0]) if ch else 0
logging.info('Parsed html data')
if s['score'] < 8 and s['num_comments'] < 6:
return False
@ -122,7 +128,7 @@ if __name__ == '__main__':
#print(self_post)
#li_comment = story('gqx')
#print(li_comment)
broken = story('n03')
broken = story('q4y')
print(broken)
# make sure there's no self-reference

@ -184,7 +184,7 @@ def feed_thread():
database.del_ref(item['ref'])
logging.info('Removed ref {}'.format(item['ref']))
gevent.sleep(6)
gevent.sleep(60)
news_index += 1
if news_index == FEED_LENGTH: news_index = 0

Loading…
Cancel
Save