diff --git a/apiserver/.gitignore b/apiserver/.gitignore index dde3e1e..8f7cd99 100644 --- a/apiserver/.gitignore +++ b/apiserver/.gitignore @@ -109,3 +109,4 @@ praw.ini data.db data.db.bak data/archive/* +qotnews.sqlite diff --git a/apiserver/database.py b/apiserver/database.py index ef31f26..826baa4 100644 --- a/apiserver/database.py +++ b/apiserver/database.py @@ -3,6 +3,7 @@ import json from sqlalchemy import create_engine, Column, String, ForeignKey, Integer from sqlalchemy.ext.declarative import declarative_base from sqlalchemy.orm import sessionmaker +from sqlalchemy.exc import IntegrityError engine = create_engine('sqlite:///data/qotnews.sqlite') Session = sessionmaker(bind=engine) @@ -15,6 +16,8 @@ class Story(Base): sid = Column(String(16), primary_key=True) meta_json = Column(String) full_json = Column(String) + title = Column(String) + date = Column(Integer) class Reflist(Base): __tablename__ = 'reflist' @@ -26,24 +29,26 @@ class Reflist(Base): def init(): Base.metadata.create_all(engine) -def get_meta(sid): +def get_story(sid): session = Session() - return session.query(Story).get(sid).meta_json - -def get_full(sid): - session = Session() - return session.query(Story).get(sid).full_json + return session.query(Story).get(sid) def put_story(story): full_json = json.dumps(story) - story.pop('text') - story.pop('comments') + story.pop('text', None) + story.pop('comments', None) meta_json = json.dumps(story) try: session = Session() - s = Story(sid=story['id'], full_json=full_json, meta_json=meta_json) + s = Story( + sid=story['id'], + full_json=full_json, + meta_json=meta_json, + title=story.get('title', None), + date=story.get('date', None), + ) session.merge(s) session.commit() except: @@ -52,11 +57,24 @@ def put_story(story): finally: session.close() +def search(q): + session = Session() + return session.query(Story).filter(Story.title.contains(q)) + def get_reflist(amount): session = Session() q = session.query(Reflist).order_by(Reflist.rid.desc()).limit(amount) return [dict(ref=x.ref, sid=x.sid) for x in q.all()] +def get_stories(amount): + session = Session() + q = session.query(Reflist, Story.meta_json).\ + order_by(Reflist.rid.desc()).\ + join(Story).\ + filter(Story.title != None).\ + limit(amount) + return [x[1] for x in q] + def put_ref(ref, sid): try: session = Session() @@ -69,5 +87,18 @@ def put_ref(ref, sid): finally: session.close() +def del_ref(ref): + try: + session = Session() + session.query(Reflist).filter(Reflist.ref==ref).delete() + session.commit() + except: + session.rollback() + raise + finally: + session.close() + if __name__ == '__main__': init() + + print(get_stories(5)) diff --git a/apiserver/feeds/hackernews.py b/apiserver/feeds/hackernews.py index 3cdc752..8d6fe5b 100644 --- a/apiserver/feeds/hackernews.py +++ b/apiserver/feeds/hackernews.py @@ -30,7 +30,7 @@ def api(route, ref=None): return False def feed(): - return api(API_TOPSTORIES) or [] + return [str(x) for x in api(API_TOPSTORIES)] or [] def comment(i): if 'author' not in i: @@ -77,6 +77,6 @@ def story(ref): # scratchpad so I can quickly develop the parser if __name__ == '__main__': - #print(feed()) + print(feed()) #print(story(20763961)) - print(story(20802050)) + #print(story(20802050)) diff --git a/apiserver/migrate-whoosh-to-sqlite.py b/apiserver/migrate-whoosh-to-sqlite.py index f793ce3..430d62b 100644 --- a/apiserver/migrate-whoosh-to-sqlite.py +++ b/apiserver/migrate-whoosh-to-sqlite.py @@ -8,14 +8,13 @@ archive.init() with archive.ix.searcher() as searcher: for docnum in searcher.document_numbers(): try: - if docnum > 500: - break + #if docnum > 500: + # break print('docnum', docnum) res = searcher.stored_fields(docnum) print('id', res['id']) database.put_story(res['story']) - database.put_ref(res['story']['ref'], res['id']) print() except BaseException as e: print('skipping', docnum) diff --git a/apiserver/server.py b/apiserver/server.py index fabddcd..64efa2a 100644 --- a/apiserver/server.py +++ b/apiserver/server.py @@ -4,13 +4,13 @@ logging.basicConfig( level=logging.INFO) import copy +import json import threading import traceback import time -import shelve from urllib.parse import urlparse, parse_qs -import archive +import database import feed from utils import gen_rand_id @@ -24,85 +24,40 @@ from gevent.pywsgi import WSGIServer monkey.patch_all() -archive.init() - -CACHE_LENGTH = 150 -DATA_FILE = 'data/data' +database.init() +FEED_LENGTH = 75 news_index = 0 -with shelve.open(DATA_FILE) as db: - logging.info('Reading caches from disk...') - news_list = db.get('news_list', []) - news_ref_to_id = db.get('news_ref_to_id', {}) - news_cache = db.get('news_cache', {}) - - # clean cache if broken - try: - for ref in news_list: - nid = news_ref_to_id[ref] - _ = news_cache[nid] - except KeyError as e: - ekey = str(e) - logging.error('Unable to find key {}. Trying to remove...'.format(ekey)) - if ekey in news_cache: - news_cache.remove(ekey) - if ekey in news_list: - news_list.remove(ekey) - if ekey in news_ref_to_id: - news_ref_to_id.remove(ekey) - -def get_story(sid): - if sid in news_cache: - return news_cache[sid] - else: - return archive.get_story(sid) - def new_id(): nid = gen_rand_id() - while nid in news_cache or archive.get_story(nid): + while database.get_story(nid): nid = gen_rand_id() return nid -def remove_ref(old_ref): - while old_ref in news_list: - news_list.remove(old_ref) - old_story = news_cache.pop(news_ref_to_id[old_ref]) - old_id = news_ref_to_id.pop(old_ref) - logging.info('Removed ref {} id {}.'.format(old_ref, old_id)) - build_folder = '../webclient/build' flask_app = Flask(__name__, template_folder=build_folder, static_folder=build_folder, static_url_path='') cors = CORS(flask_app) @flask_app.route('/api') def api(): - try: - front_page = [news_cache[news_ref_to_id[ref]] for ref in news_list] - except KeyError as e: - ekey = str(e) - logging.error('Unable to find key {}. Trying to remove...'.format(ekey)) - if ekey in news_cache: - news_cache.remove(ekey) - if ekey in news_list: - news_list.remove(ekey) - - front_page = [copy.copy(x) for x in front_page if 'title' in x and x['title']] - front_page = front_page[:60] - for story in front_page: - story.pop('text', None) - story.pop('comments', None) - - return {'stories': front_page} + stories = database.get_stories(FEED_LENGTH) + # hacky nested json + res = Response('{"stories":[' + ','.join(stories) + ']}') + res.headers['content-type'] = 'application/json' + return res @flask_app.route('/api/search', strict_slashes=False) def search(): - search = request.args.get('q', '') - if len(search) >= 3: - res = archive.search(search) + q = request.args.get('q', '') + if len(q) >= 3: + results = [x.meta_json for x in database.search(q)] else: - res = [] - return {'results': res} + results = [] + # hacky nested json + res = Response('{"results":[' + ','.join(results) + ']}') + res.headers['content-type'] = 'application/json' + return res @flask_app.route('/api/submit', methods=['POST'], strict_slashes=False) def submit(): @@ -124,14 +79,14 @@ def submit(): source = 'manual' ref = url - news_story = dict(id=nid, ref=ref, source=source) - news_cache[nid] = news_story - valid = feed.update_story(news_story, is_manual=True) + # TODO: return existing refs + + story = dict(id=nid, ref=ref, source=source) + valid = feed.update_story(story, is_manual=True) if valid: - archive.update(news_story) + database.put_story(story) return {'nid': nid} else: - news_cache.pop(nid, '') raise Exception('Invalid article') except BaseException as e: @@ -142,8 +97,14 @@ def submit(): @flask_app.route('/api/') def story(sid): - story = get_story(sid) - return dict(story=story) if story else abort(404) + story = database.get_story(sid) + if story: + # hacky nested json + res = Response('{"story":' + story.full_json + '}') + res.headers['content-type'] = 'application/json' + return res + else: + return abort(404) @flask_app.route('/') @flask_app.route('/search') @@ -161,8 +122,9 @@ def static_story(sid): except NotFound: pass - story = get_story(sid) + story = database.get_story(sid) if not story: return abort(404) + story = json.loads(story.full_json) score = story['score'] num_comments = story['num_comments'] @@ -179,58 +141,43 @@ def static_story(sid): url=url, description=description) -http_server = WSGIServer(('', 33842), flask_app) +http_server = WSGIServer(('', 43842), flask_app) def feed_thread(): global news_index try: while True: + ref_list = database.get_reflist(FEED_LENGTH) + # onboard new stories if news_index == 0: - feed_list = feed.list() - new_items = [(ref, source) for ref, source in feed_list if ref not in news_list] - for ref, source in new_items: - news_list.insert(0, ref) - nid = new_id() - news_ref_to_id[ref] = nid - news_cache[nid] = dict(id=nid, ref=ref, source=source) - - if len(new_items): - logging.info('Added {} new refs.'.format(len(new_items))) - - # drop old ones - while len(news_list) > CACHE_LENGTH: - old_ref = news_list[-1] - remove_ref(old_ref) + for ref, source in feed.list(): + try: + nid = new_id() + database.put_ref(ref, nid) + database.put_story(dict(id=nid, ref=ref, source=source)) + logging.info('Added ref ' + ref) + except database.IntegrityError: + continue # update current stories - if news_index < len(news_list): - try: - update_ref = news_list[news_index] - update_id = news_ref_to_id[update_ref] - news_story = news_cache[update_id] - except KeyError as e: - ekey = str(e) - logging.error('Unable to find key {}. Trying to remove...'.format(ekey)) - if ekey in news_cache: - news_cache.remove(ekey) - if ekey in news_list: - news_list.remove(ekey) - if ekey in news_ref_to_id: - news_ref_to_id.remove(ekey) - valid = feed.update_story(news_story) + if news_index < len(ref_list): + update_ref = ref_list[news_index]['ref'] + update_sid = ref_list[news_index]['sid'] + story_json = database.get_story(update_sid).full_json + story = json.loads(story_json) + valid = feed.update_story(story) if valid: - archive.update(news_story) + database.put_story(story) else: - remove_ref(update_ref) - else: - logging.info('Skipping update - no story #' + str(news_index+1)) + database.del_ref(update_ref) + logging.info('Removed ref {}'.format(update_ref)) gevent.sleep(6) news_index += 1 - if news_index == CACHE_LENGTH: news_index = 0 + if news_index == FEED_LENGTH: news_index = 0 except KeyboardInterrupt: logging.info('Ending feed thread...') @@ -246,9 +193,3 @@ try: http_server.serve_forever() except KeyboardInterrupt: logging.info('Exiting...') -finally: - with shelve.open(DATA_FILE) as db: - logging.info('Writing caches to disk...') - db['news_list'] = news_list - db['news_ref_to_id'] = news_ref_to_id - db['news_cache'] = news_cache