292 lines
10 KiB
Python
292 lines
10 KiB
Python
import asyncio
|
|
import json
|
|
import time
|
|
|
|
from sqlalchemy import select, or_, and_, text, desc
|
|
from sqlalchemy.orm import aliased
|
|
from sqlalchemy_searchable import search
|
|
|
|
from orm.author import Author, AuthorFollower
|
|
from orm.shout import ShoutAuthor, ShoutTopic
|
|
from orm.topic import Topic
|
|
from resolvers.stat import get_with_stat, author_follows_authors, author_follows_topics
|
|
from services.cache import set_author_cache, update_author_followers_cache
|
|
from services.auth import login_required
|
|
from services.db import local_session
|
|
from services.encoders import CustomJSONEncoder
|
|
from services.rediscache import redis
|
|
from services.schema import mutation, query
|
|
from services.logger import root_logger as logger
|
|
|
|
|
|
@mutation.field('update_author')
|
|
@login_required
|
|
async def update_author(_, info, profile):
|
|
user_id = info.context['user_id']
|
|
with local_session() as session:
|
|
author = session.query(Author).where(Author.user == user_id).first()
|
|
Author.update(author, profile)
|
|
session.add(author)
|
|
session.commit()
|
|
return {'error': None, 'author': author}
|
|
|
|
|
|
@query.field('get_authors_all')
|
|
def get_authors_all(_, _info):
|
|
with local_session() as session:
|
|
authors = session.query(Author).all()
|
|
return authors
|
|
|
|
|
|
@query.field('get_author')
|
|
async def get_author(_, _info, slug='', author_id=None):
|
|
author = None
|
|
try:
|
|
if slug:
|
|
q = select(Author).select_from(Author).filter(Author.slug == slug)
|
|
result = get_with_stat(q)
|
|
if result:
|
|
[author] = result
|
|
author_id = author.id
|
|
|
|
if author_id:
|
|
cache = await redis.execute('GET', f'id:{author_id}:author')
|
|
logger.debug(f'result from cache: {cache}')
|
|
q = select(Author).where(Author.id == author_id)
|
|
author_dict = None
|
|
if cache:
|
|
author_dict = json.loads(cache)
|
|
else:
|
|
result = get_with_stat(q)
|
|
if result:
|
|
[author] = result
|
|
author_dict = author.dict()
|
|
logger.debug(f'author to be stored: {author_dict}')
|
|
if author:
|
|
await set_author_cache(author_dict)
|
|
return author_dict
|
|
except Exception as exc:
|
|
import traceback
|
|
|
|
logger.error(exc)
|
|
exc = traceback.format_exc()
|
|
logger.error(exc)
|
|
return
|
|
# {"slug": "anonymous", "id": 1, "name": "Аноним", "bio": "Неизвестно кто"}
|
|
|
|
|
|
async def get_author_by_user_id(user_id: str):
|
|
logger.info(f'getting author id for {user_id}')
|
|
redis_key = f'user:{user_id}:author'
|
|
author = None
|
|
try:
|
|
res = await redis.execute('GET', redis_key)
|
|
if isinstance(res, str):
|
|
author = json.loads(res)
|
|
author_id = author.get('id')
|
|
author_slug = author.get('slug')
|
|
if author_id:
|
|
logger.debug(f'got author @{author_slug} #{author_id} cached')
|
|
return author
|
|
|
|
q = select(Author).filter(Author.user == user_id)
|
|
result = get_with_stat(q)
|
|
if result:
|
|
[author] = result
|
|
await set_author_cache(author.dict())
|
|
except Exception as exc:
|
|
import traceback
|
|
|
|
traceback.print_exc()
|
|
logger.error(exc)
|
|
return author
|
|
|
|
|
|
@query.field('get_author_id')
|
|
async def get_author_id(_, _info, user: str):
|
|
return await get_author_by_user_id(user)
|
|
|
|
|
|
@query.field('load_authors_by')
|
|
def load_authors_by(_, _info, by, limit, offset):
|
|
logger.debug(f'loading authors by {by}')
|
|
q = select(Author)
|
|
if by.get('slug'):
|
|
q = q.filter(Author.slug.ilike(f"%{by['slug']}%"))
|
|
elif by.get('name'):
|
|
q = q.filter(Author.name.ilike(f"%{by['name']}%"))
|
|
elif by.get('topic'):
|
|
q = (
|
|
q.join(ShoutAuthor)
|
|
.join(ShoutTopic)
|
|
.join(Topic)
|
|
.where(Topic.slug == str(by['topic']))
|
|
)
|
|
|
|
if by.get('last_seen'): # in unix time
|
|
before = int(time.time()) - by['last_seen']
|
|
q = q.filter(Author.last_seen > before)
|
|
elif by.get('created_at'): # in unix time
|
|
before = int(time.time()) - by['created_at']
|
|
q = q.filter(Author.created_at > before)
|
|
|
|
order = by.get('order')
|
|
if order in ['likes', 'shouts', 'followers']:
|
|
q = q.order_by(desc(text(f'{order}_stat')))
|
|
|
|
# q = q.distinct()
|
|
q = q.limit(limit).offset(offset)
|
|
|
|
authors = get_with_stat(q)
|
|
|
|
return authors
|
|
|
|
|
|
@query.field('get_author_follows')
|
|
async def get_author_follows(_, _info, slug='', user=None, author_id=None):
|
|
with (local_session() as session):
|
|
if user or slug:
|
|
author = session.query(Author).filter(or_(Author.user == user, Author.slug == slug)).first()
|
|
if author:
|
|
author_id = author.id
|
|
if author_id:
|
|
rkey = f'author:{author_id}:follows-authors'
|
|
logger.debug(f'getting {author_id} follows authors')
|
|
cached = await redis.execute('GET', rkey)
|
|
if not cached:
|
|
authors = author_follows_authors(author_id)
|
|
prepared = [author.dict() for author in authors]
|
|
await redis.execute('SET', rkey, json.dumps(prepared, cls=CustomJSONEncoder))
|
|
else:
|
|
authors = json.loads(cached)
|
|
|
|
rkey = f'author:{author_id}:follows-topics'
|
|
cached = await redis.execute('GET', rkey)
|
|
if cached:
|
|
topics = json.loads(cached)
|
|
if not cached:
|
|
topics = author_follows_topics(author_id)
|
|
prepared = [topic.dict() for topic in topics]
|
|
await redis.execute('SET', rkey, json.dumps(prepared, cls=CustomJSONEncoder))
|
|
return {
|
|
'topics': topics,
|
|
'authors': authors,
|
|
'communities': [
|
|
{'id': 1, 'name': 'Дискурс', 'slug': 'discours', 'pic': ''}
|
|
],
|
|
}
|
|
else:
|
|
raise ValueError('Author not found')
|
|
|
|
|
|
@query.field('get_author_follows_topics')
|
|
async def get_author_follows_topics(_, _info, slug='', user=None, author_id=None):
|
|
with local_session() as session:
|
|
if user or slug:
|
|
author_id_result = (
|
|
session.query(Author.id)
|
|
.filter(or_(Author.user == user, Author.slug == slug))
|
|
.first()
|
|
)
|
|
author_id = author_id_result[0] if author_id_result else None
|
|
if author_id:
|
|
logger.debug(f'getting {author_id} follows topics')
|
|
rkey = f'author:{author_id}:follows-topics'
|
|
cached = await redis.execute('GET', rkey)
|
|
topics = json.loads(cached) if cached else author_follows_topics(author_id)
|
|
if not cached:
|
|
prepared = [topic.dict() for topic in topics]
|
|
await redis.execute('SET', rkey, json.dumps(prepared, cls=CustomJSONEncoder))
|
|
return topics
|
|
else:
|
|
raise ValueError('Author not found')
|
|
|
|
|
|
@query.field('get_author_follows_authors')
|
|
async def get_author_follows_authors(_, _info, slug='', user=None, author_id=None):
|
|
with local_session() as session:
|
|
if user or slug:
|
|
author_id_result = (
|
|
session.query(Author.id)
|
|
.filter(or_(Author.user == user, Author.slug == slug))
|
|
.first()
|
|
)
|
|
author_id = author_id_result[0] if author_id_result else None
|
|
if author_id:
|
|
logger.debug(f'getting {author_id} follows authors')
|
|
rkey = f'author:{author_id}:follows-authors'
|
|
cached = await redis.execute('GET', rkey)
|
|
authors = (
|
|
json.loads(cached) if cached else author_follows_authors(author_id)
|
|
)
|
|
if not cached:
|
|
prepared = [author.dict() for author in authors]
|
|
await redis.execute('SET', rkey, json.dumps(prepared, cls=CustomJSONEncoder))
|
|
return authors
|
|
else:
|
|
raise ValueError('Author not found')
|
|
|
|
|
|
def create_author(user_id: str, slug: str, name: str = ''):
|
|
with local_session() as session:
|
|
try:
|
|
author = None
|
|
if user_id:
|
|
author = session.query(Author).filter(Author.user == user_id).first()
|
|
elif slug:
|
|
author = session.query(Author).filter(Author.slug == slug).first()
|
|
if not author:
|
|
new_author = Author(user=user_id, slug=slug, name=name)
|
|
session.add(new_author)
|
|
session.commit()
|
|
logger.info(f'author created by webhook {new_author.dict()}')
|
|
except Exception as exc:
|
|
logger.debug(exc)
|
|
|
|
|
|
@query.field('get_author_followers')
|
|
async def get_author_followers(_, _info, slug: str):
|
|
logger.debug(f'getting followers for @{slug}')
|
|
try:
|
|
with local_session() as session:
|
|
author_alias = aliased(Author)
|
|
author_id = (
|
|
session.query(author_alias.id)
|
|
.filter(author_alias.slug == slug)
|
|
.scalar()
|
|
)
|
|
if author_id:
|
|
cached = await redis.execute('GET', f'author:{author_id}:followers')
|
|
if not cached:
|
|
author_follower_alias = aliased(AuthorFollower, name='af')
|
|
q = select(Author).join(
|
|
author_follower_alias,
|
|
and_(
|
|
author_follower_alias.author == author_id,
|
|
author_follower_alias.follower == Author.id,
|
|
),
|
|
)
|
|
results = get_with_stat(q)
|
|
_ = asyncio.create_task(
|
|
update_author_followers_cache(
|
|
author_id, [x.dict() for x in results]
|
|
)
|
|
)
|
|
logger.debug(f'@{slug} cache updated with {len(results)} followers')
|
|
return results
|
|
else:
|
|
logger.debug(f'@{slug} got followers cached')
|
|
return json.loads(cached)
|
|
except Exception as exc:
|
|
import traceback
|
|
|
|
logger.error(exc)
|
|
logger.error(traceback.format_exc())
|
|
return []
|
|
|
|
|
|
@query.field('search_authors')
|
|
def search_authors(_, _info, what: str):
|
|
q = search(select(Author), what)
|
|
return get_with_stat(q)
|