configured isort, black, flake8
This commit is contained in:
parent
17c29c7f4f
commit
441bcc1e90
6
.flake8
6
.flake8
|
@ -1,6 +0,0 @@
|
||||||
[flake8]
|
|
||||||
ignore = E203,W504,W191,W503
|
|
||||||
exclude = .git,__pycache__,orm/rbac.py
|
|
||||||
max-complexity = 10
|
|
||||||
max-line-length = 108
|
|
||||||
indent-string = ' '
|
|
16
.github/workflows/checks.yml
vendored
Normal file
16
.github/workflows/checks.yml
vendored
Normal file
|
@ -0,0 +1,16 @@
|
||||||
|
name: Checks
|
||||||
|
on: [pull_request]
|
||||||
|
|
||||||
|
jobs:
|
||||||
|
build:
|
||||||
|
runs-on: ubuntu-latest
|
||||||
|
name: Checks
|
||||||
|
steps:
|
||||||
|
- uses: actions/checkout@v2
|
||||||
|
- uses: actions/setup-python@v2
|
||||||
|
with:
|
||||||
|
python-version: 3.10
|
||||||
|
- run: pip install --upgrade pip
|
||||||
|
- run: pip install -r requirements.txt
|
||||||
|
- run: pip install -r requirements-dev.txt
|
||||||
|
- run: check.sh
|
|
@ -6,11 +6,11 @@ exclude: |
|
||||||
)
|
)
|
||||||
|
|
||||||
default_language_version:
|
default_language_version:
|
||||||
python: python3.8
|
python: python3.10
|
||||||
|
|
||||||
repos:
|
repos:
|
||||||
- repo: https://github.com/pre-commit/pre-commit-hooks
|
- repo: https://github.com/pre-commit/pre-commit-hooks
|
||||||
rev: v3.2.0
|
rev: v4.5.0
|
||||||
hooks:
|
hooks:
|
||||||
- id: check-added-large-files
|
- id: check-added-large-files
|
||||||
- id: check-case-conflict
|
- id: check-case-conflict
|
||||||
|
@ -21,24 +21,24 @@ repos:
|
||||||
- id: check-yaml
|
- id: check-yaml
|
||||||
- id: end-of-file-fixer
|
- id: end-of-file-fixer
|
||||||
- id: trailing-whitespace
|
- id: trailing-whitespace
|
||||||
|
- id: requirements-txt-fixer
|
||||||
|
|
||||||
- repo: https://github.com/timothycrosley/isort
|
- repo: https://github.com/timothycrosley/isort
|
||||||
rev: 5.5.3
|
rev: 5.12.0
|
||||||
hooks:
|
hooks:
|
||||||
- id: isort
|
- id: isort
|
||||||
|
|
||||||
- repo: https://github.com/ambv/black
|
- repo: https://github.com/ambv/black
|
||||||
rev: 20.8b1
|
rev: 23.10.1
|
||||||
hooks:
|
hooks:
|
||||||
- id: black
|
- id: black
|
||||||
args:
|
|
||||||
- --line-length=100
|
|
||||||
- --skip-string-normalization
|
|
||||||
|
|
||||||
- repo: https://gitlab.com/pycqa/flake8
|
- repo: https://github.com/PyCQA/flake8
|
||||||
rev: 3.8.3
|
rev: 6.1.0
|
||||||
hooks:
|
hooks:
|
||||||
- id: flake8
|
- id: flake8
|
||||||
args:
|
|
||||||
- --max-line-length=100
|
# - repo: https://github.com/python/mypy
|
||||||
- --disable=protected-access
|
# rev: v1.6.1
|
||||||
|
# hooks:
|
||||||
|
# - id: mypy
|
||||||
|
|
|
@ -42,4 +42,3 @@ Put the header 'Authorization' with token from signIn query or registerUser muta
|
||||||
# How to debug Ackee
|
# How to debug Ackee
|
||||||
|
|
||||||
Set ACKEE_TOKEN var
|
Set ACKEE_TOKEN var
|
||||||
|
|
||||||
|
|
|
@ -1,75 +0,0 @@
|
||||||
import re
|
|
||||||
import nltk
|
|
||||||
from bs4 import BeautifulSoup
|
|
||||||
from nltk.corpus import stopwords
|
|
||||||
from pymystem3 import Mystem
|
|
||||||
from string import punctuation
|
|
||||||
from transformers import BertTokenizer
|
|
||||||
|
|
||||||
nltk.download("stopwords")
|
|
||||||
|
|
||||||
|
|
||||||
def get_clear_text(text):
|
|
||||||
soup = BeautifulSoup(text, 'html.parser')
|
|
||||||
|
|
||||||
# extract the plain text from the HTML document without tags
|
|
||||||
clear_text = ''
|
|
||||||
for tag in soup.find_all():
|
|
||||||
clear_text += tag.string or ''
|
|
||||||
|
|
||||||
clear_text = re.sub(pattern='[\u202F\u00A0\n]+', repl=' ', string=clear_text)
|
|
||||||
|
|
||||||
# only words
|
|
||||||
clear_text = re.sub(pattern='[^A-ZА-ЯЁ -]', repl='', string=clear_text, flags=re.IGNORECASE)
|
|
||||||
|
|
||||||
clear_text = re.sub(pattern='\s+', repl=' ', string=clear_text)
|
|
||||||
|
|
||||||
clear_text = clear_text.lower()
|
|
||||||
|
|
||||||
mystem = Mystem()
|
|
||||||
russian_stopwords = stopwords.words("russian")
|
|
||||||
|
|
||||||
tokens = mystem.lemmatize(clear_text)
|
|
||||||
tokens = [token for token in tokens if token not in russian_stopwords \
|
|
||||||
and token != " " \
|
|
||||||
and token.strip() not in punctuation]
|
|
||||||
|
|
||||||
clear_text = " ".join(tokens)
|
|
||||||
|
|
||||||
return clear_text
|
|
||||||
|
|
||||||
|
|
||||||
# if __name__ == '__main__':
|
|
||||||
#
|
|
||||||
# # initialize the tokenizer with the pre-trained BERT model and vocabulary
|
|
||||||
# tokenizer = BertTokenizer.from_pretrained('bert-base-multilingual-cased')
|
|
||||||
#
|
|
||||||
# # split each text into smaller segments of maximum length 512
|
|
||||||
# max_length = 512
|
|
||||||
# segmented_texts = []
|
|
||||||
# for text in [clear_text1, clear_text2]:
|
|
||||||
# segmented_text = []
|
|
||||||
# for i in range(0, len(text), max_length):
|
|
||||||
# segment = text[i:i+max_length]
|
|
||||||
# segmented_text.append(segment)
|
|
||||||
# segmented_texts.append(segmented_text)
|
|
||||||
#
|
|
||||||
# # tokenize each segment using the BERT tokenizer
|
|
||||||
# tokenized_texts = []
|
|
||||||
# for segmented_text in segmented_texts:
|
|
||||||
# tokenized_text = []
|
|
||||||
# for segment in segmented_text:
|
|
||||||
# segment_tokens = tokenizer.tokenize(segment)
|
|
||||||
# segment_tokens = ['[CLS]'] + segment_tokens + ['[SEP]']
|
|
||||||
# tokenized_text.append(segment_tokens)
|
|
||||||
# tokenized_texts.append(tokenized_text)
|
|
||||||
#
|
|
||||||
# input_ids = []
|
|
||||||
# for tokenized_text in tokenized_texts:
|
|
||||||
# input_id = []
|
|
||||||
# for segment_tokens in tokenized_text:
|
|
||||||
# segment_id = tokenizer.convert_tokens_to_ids(segment_tokens)
|
|
||||||
# input_id.append(segment_id)
|
|
||||||
# input_ids.append(input_id)
|
|
||||||
#
|
|
||||||
# print(input_ids)
|
|
|
@ -1,10 +1,9 @@
|
||||||
from logging.config import fileConfig
|
from logging.config import fileConfig
|
||||||
|
|
||||||
from sqlalchemy import engine_from_config
|
from sqlalchemy import engine_from_config, pool
|
||||||
from sqlalchemy import pool
|
|
||||||
|
|
||||||
from alembic import context
|
from alembic import context
|
||||||
|
from base.orm import Base
|
||||||
from settings import DB_URL
|
from settings import DB_URL
|
||||||
|
|
||||||
# this is the Alembic Config object, which provides
|
# this is the Alembic Config object, which provides
|
||||||
|
@ -19,7 +18,6 @@ config.set_section_option(config.config_ini_section, "DB_URL", DB_URL)
|
||||||
if config.config_file_name is not None:
|
if config.config_file_name is not None:
|
||||||
fileConfig(config.config_file_name)
|
fileConfig(config.config_file_name)
|
||||||
|
|
||||||
from base.orm import Base
|
|
||||||
target_metadata = [Base.metadata]
|
target_metadata = [Base.metadata]
|
||||||
|
|
||||||
# other values from the config, defined by the needs of env.py,
|
# other values from the config, defined by the needs of env.py,
|
||||||
|
@ -66,9 +64,7 @@ def run_migrations_online() -> None:
|
||||||
)
|
)
|
||||||
|
|
||||||
with connectable.connect() as connection:
|
with connectable.connect() as connection:
|
||||||
context.configure(
|
context.configure(connection=connection, target_metadata=target_metadata)
|
||||||
connection=connection, target_metadata=target_metadata
|
|
||||||
)
|
|
||||||
|
|
||||||
with context.begin_transaction():
|
with context.begin_transaction():
|
||||||
context.run_migrations()
|
context.run_migrations()
|
||||||
|
|
|
@ -7,12 +7,12 @@ Create Date: 2023-08-19 01:37:57.031933
|
||||||
"""
|
"""
|
||||||
from typing import Sequence, Union
|
from typing import Sequence, Union
|
||||||
|
|
||||||
from alembic import op
|
# import sqlalchemy as sa
|
||||||
import sqlalchemy as sa
|
|
||||||
|
|
||||||
|
# from alembic import op
|
||||||
|
|
||||||
# revision identifiers, used by Alembic.
|
# revision identifiers, used by Alembic.
|
||||||
revision: str = 'fe943b098418'
|
revision: str = "fe943b098418"
|
||||||
down_revision: Union[str, None] = None
|
down_revision: Union[str, None] = None
|
||||||
branch_labels: Union[str, Sequence[str], None] = None
|
branch_labels: Union[str, Sequence[str], None] = None
|
||||||
depends_on: Union[str, Sequence[str], None] = None
|
depends_on: Union[str, Sequence[str], None] = None
|
||||||
|
|
|
@ -2,75 +2,71 @@ from functools import wraps
|
||||||
from typing import Optional, Tuple
|
from typing import Optional, Tuple
|
||||||
|
|
||||||
from graphql.type import GraphQLResolveInfo
|
from graphql.type import GraphQLResolveInfo
|
||||||
from sqlalchemy.orm import joinedload, exc
|
from sqlalchemy.orm import exc, joinedload
|
||||||
from starlette.authentication import AuthenticationBackend
|
from starlette.authentication import AuthenticationBackend
|
||||||
from starlette.requests import HTTPConnection
|
from starlette.requests import HTTPConnection
|
||||||
|
|
||||||
from auth.credentials import AuthCredentials, AuthUser
|
from auth.credentials import AuthCredentials, AuthUser
|
||||||
from base.orm import local_session
|
|
||||||
from orm.user import User, Role
|
|
||||||
|
|
||||||
from settings import SESSION_TOKEN_HEADER
|
|
||||||
from auth.tokenstorage import SessionToken
|
from auth.tokenstorage import SessionToken
|
||||||
from base.exceptions import OperationNotAllowed
|
from base.exceptions import OperationNotAllowed
|
||||||
|
from base.orm import local_session
|
||||||
|
from orm.user import Role, User
|
||||||
|
from settings import SESSION_TOKEN_HEADER
|
||||||
|
|
||||||
|
|
||||||
class JWTAuthenticate(AuthenticationBackend):
|
class JWTAuthenticate(AuthenticationBackend):
|
||||||
async def authenticate(
|
async def authenticate(
|
||||||
self, request: HTTPConnection
|
self, request: HTTPConnection
|
||||||
) -> Optional[Tuple[AuthCredentials, AuthUser]]:
|
) -> Optional[Tuple[AuthCredentials, AuthUser]]:
|
||||||
|
|
||||||
if SESSION_TOKEN_HEADER not in request.headers:
|
if SESSION_TOKEN_HEADER not in request.headers:
|
||||||
return AuthCredentials(scopes={}), AuthUser(user_id=None, username='')
|
return AuthCredentials(scopes={}), AuthUser(user_id=None, username="")
|
||||||
|
|
||||||
token = request.headers.get(SESSION_TOKEN_HEADER)
|
token = request.headers.get(SESSION_TOKEN_HEADER)
|
||||||
if not token:
|
if not token:
|
||||||
print("[auth.authenticate] no token in header %s" % SESSION_TOKEN_HEADER)
|
print("[auth.authenticate] no token in header %s" % SESSION_TOKEN_HEADER)
|
||||||
return AuthCredentials(scopes={}, error_message=str("no token")), AuthUser(
|
return AuthCredentials(scopes={}, error_message=str("no token")), AuthUser(
|
||||||
user_id=None, username=''
|
user_id=None, username=""
|
||||||
)
|
)
|
||||||
|
|
||||||
if len(token.split('.')) > 1:
|
if len(token.split(".")) > 1:
|
||||||
payload = await SessionToken.verify(token)
|
payload = await SessionToken.verify(token)
|
||||||
|
|
||||||
with local_session() as session:
|
with local_session() as session:
|
||||||
try:
|
try:
|
||||||
user = (
|
user = (
|
||||||
session.query(User).options(
|
session.query(User)
|
||||||
|
.options(
|
||||||
joinedload(User.roles).options(joinedload(Role.permissions)),
|
joinedload(User.roles).options(joinedload(Role.permissions)),
|
||||||
joinedload(User.ratings)
|
joinedload(User.ratings),
|
||||||
).filter(
|
)
|
||||||
User.id == payload.user_id
|
.filter(User.id == payload.user_id)
|
||||||
).one()
|
.one()
|
||||||
)
|
)
|
||||||
|
|
||||||
scopes = {} # TODO: integrate await user.get_permission()
|
scopes = {} # TODO: integrate await user.get_permission()
|
||||||
|
|
||||||
return (
|
return (
|
||||||
AuthCredentials(
|
AuthCredentials(user_id=payload.user_id, scopes=scopes, logged_in=True),
|
||||||
user_id=payload.user_id,
|
AuthUser(user_id=user.id, username=""),
|
||||||
scopes=scopes,
|
|
||||||
logged_in=True
|
|
||||||
),
|
|
||||||
AuthUser(user_id=user.id, username=''),
|
|
||||||
)
|
)
|
||||||
except exc.NoResultFound:
|
except exc.NoResultFound:
|
||||||
pass
|
pass
|
||||||
|
|
||||||
return AuthCredentials(scopes={}, error_message=str('Invalid token')), AuthUser(user_id=None, username='')
|
return AuthCredentials(scopes={}, error_message=str("Invalid token")), AuthUser(
|
||||||
|
user_id=None, username=""
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
def login_required(func):
|
def login_required(func):
|
||||||
@wraps(func)
|
@wraps(func)
|
||||||
async def wrap(parent, info: GraphQLResolveInfo, *args, **kwargs):
|
async def wrap(parent, info: GraphQLResolveInfo, *args, **kwargs):
|
||||||
# print('[auth.authenticate] login required for %r with info %r' % (func, info)) # debug only
|
# debug only
|
||||||
|
# print('[auth.authenticate] login required for %r with info %r' % (func, info))
|
||||||
auth: AuthCredentials = info.context["request"].auth
|
auth: AuthCredentials = info.context["request"].auth
|
||||||
# print(auth)
|
# print(auth)
|
||||||
if not auth or not auth.logged_in:
|
if not auth or not auth.logged_in:
|
||||||
# raise Unauthorized(auth.error_message or "Please login")
|
# raise Unauthorized(auth.error_message or "Please login")
|
||||||
return {
|
return {"error": "Please login first"}
|
||||||
"error": "Please login first"
|
|
||||||
}
|
|
||||||
return await func(parent, info, *args, **kwargs)
|
return await func(parent, info, *args, **kwargs)
|
||||||
|
|
||||||
return wrap
|
return wrap
|
||||||
|
@ -79,7 +75,9 @@ def login_required(func):
|
||||||
def permission_required(resource, operation, func):
|
def permission_required(resource, operation, func):
|
||||||
@wraps(func)
|
@wraps(func)
|
||||||
async def wrap(parent, info: GraphQLResolveInfo, *args, **kwargs):
|
async def wrap(parent, info: GraphQLResolveInfo, *args, **kwargs):
|
||||||
print('[auth.authenticate] permission_required for %r with info %r' % (func, info)) # debug only
|
print(
|
||||||
|
"[auth.authenticate] permission_required for %r with info %r" % (func, info)
|
||||||
|
) # debug only
|
||||||
auth: AuthCredentials = info.context["request"].auth
|
auth: AuthCredentials = info.context["request"].auth
|
||||||
if not auth.logged_in:
|
if not auth.logged_in:
|
||||||
raise OperationNotAllowed(auth.error_message or "Please login")
|
raise OperationNotAllowed(auth.error_message or "Please login")
|
||||||
|
|
|
@ -23,13 +23,11 @@ class AuthCredentials(BaseModel):
|
||||||
async def permissions(self) -> List[Permission]:
|
async def permissions(self) -> List[Permission]:
|
||||||
if self.user_id is None:
|
if self.user_id is None:
|
||||||
# raise Unauthorized("Please login first")
|
# raise Unauthorized("Please login first")
|
||||||
return {
|
return {"error": "Please login first"}
|
||||||
"error": "Please login first"
|
|
||||||
}
|
|
||||||
else:
|
else:
|
||||||
# TODO: implement permissions logix
|
# TODO: implement permissions logix
|
||||||
print(self.user_id)
|
print(self.user_id)
|
||||||
return NotImplemented()
|
return NotImplemented
|
||||||
|
|
||||||
|
|
||||||
class AuthUser(BaseModel):
|
class AuthUser(BaseModel):
|
||||||
|
@ -40,6 +38,6 @@ class AuthUser(BaseModel):
|
||||||
def is_authenticated(self) -> bool:
|
def is_authenticated(self) -> bool:
|
||||||
return self.user_id is not None
|
return self.user_id is not None
|
||||||
|
|
||||||
@property
|
# @property
|
||||||
def display_id(self) -> int:
|
# def display_id(self) -> int:
|
||||||
return self.user_id
|
# return self.user_id
|
||||||
|
|
|
@ -2,19 +2,16 @@ import requests
|
||||||
|
|
||||||
from settings import MAILGUN_API_KEY, MAILGUN_DOMAIN
|
from settings import MAILGUN_API_KEY, MAILGUN_DOMAIN
|
||||||
|
|
||||||
api_url = "https://api.mailgun.net/v3/%s/messages" % (MAILGUN_DOMAIN or 'discours.io')
|
api_url = "https://api.mailgun.net/v3/%s/messages" % (MAILGUN_DOMAIN or "discours.io")
|
||||||
noreply = "discours.io <noreply@%s>" % (MAILGUN_DOMAIN or 'discours.io')
|
noreply = "discours.io <noreply@%s>" % (MAILGUN_DOMAIN or "discours.io")
|
||||||
lang_subject = {
|
lang_subject = {"ru": "Подтверждение почты", "en": "Confirm email"}
|
||||||
"ru": "Подтверждение почты",
|
|
||||||
"en": "Confirm email"
|
|
||||||
}
|
|
||||||
|
|
||||||
|
|
||||||
async def send_auth_email(user, token, lang="ru", template="email_confirmation"):
|
async def send_auth_email(user, token, lang="ru", template="email_confirmation"):
|
||||||
try:
|
try:
|
||||||
to = "%s <%s>" % (user.name, user.email)
|
to = "%s <%s>" % (user.name, user.email)
|
||||||
if lang not in ['ru', 'en']:
|
if lang not in ["ru", "en"]:
|
||||||
lang = 'ru'
|
lang = "ru"
|
||||||
subject = lang_subject.get(lang, lang_subject["en"])
|
subject = lang_subject.get(lang, lang_subject["en"])
|
||||||
template = template + "_" + lang
|
template = template + "_" + lang
|
||||||
payload = {
|
payload = {
|
||||||
|
@ -22,16 +19,12 @@ async def send_auth_email(user, token, lang="ru", template="email_confirmation")
|
||||||
"to": to,
|
"to": to,
|
||||||
"subject": subject,
|
"subject": subject,
|
||||||
"template": template,
|
"template": template,
|
||||||
"h:X-Mailgun-Variables": "{ \"token\": \"%s\" }" % token
|
"h:X-Mailgun-Variables": '{ "token": "%s" }' % token,
|
||||||
}
|
}
|
||||||
print('[auth.email] payload: %r' % payload)
|
print("[auth.email] payload: %r" % payload)
|
||||||
# debug
|
# debug
|
||||||
# print('http://localhost:3000/?modal=auth&mode=confirm-email&token=%s' % token)
|
# print('http://localhost:3000/?modal=auth&mode=confirm-email&token=%s' % token)
|
||||||
response = requests.post(
|
response = requests.post(api_url, auth=("api", MAILGUN_API_KEY), data=payload)
|
||||||
api_url,
|
|
||||||
auth=("api", MAILGUN_API_KEY),
|
|
||||||
data=payload
|
|
||||||
)
|
|
||||||
response.raise_for_status()
|
response.raise_for_status()
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
print(e)
|
print(e)
|
||||||
|
|
|
@ -7,6 +7,7 @@ from sqlalchemy import or_
|
||||||
|
|
||||||
from auth.jwtcodec import JWTCodec
|
from auth.jwtcodec import JWTCodec
|
||||||
from auth.tokenstorage import TokenStorage
|
from auth.tokenstorage import TokenStorage
|
||||||
|
|
||||||
# from base.exceptions import InvalidPassword, InvalidToken
|
# from base.exceptions import InvalidPassword, InvalidToken
|
||||||
from base.orm import local_session
|
from base.orm import local_session
|
||||||
from orm import User
|
from orm import User
|
||||||
|
@ -34,7 +35,7 @@ class Password:
|
||||||
Verify that password hash is equal to specified hash. Hash format:
|
Verify that password hash is equal to specified hash. Hash format:
|
||||||
|
|
||||||
$2a$10$Ro0CUfOqk6cXEKf3dyaM7OhSCvnwM9s4wIX9JeLapehKK5YdLxKcm
|
$2a$10$Ro0CUfOqk6cXEKf3dyaM7OhSCvnwM9s4wIX9JeLapehKK5YdLxKcm
|
||||||
\__/\/ \____________________/\_____________________________/
|
\__/\/ \____________________/\_____________________________/ # noqa: W605
|
||||||
| | Salt Hash
|
| | Salt Hash
|
||||||
| Cost
|
| Cost
|
||||||
Version
|
Version
|
||||||
|
@ -57,14 +58,10 @@ class Identity:
|
||||||
user = User(**orm_user.dict())
|
user = User(**orm_user.dict())
|
||||||
if not user.password:
|
if not user.password:
|
||||||
# raise InvalidPassword("User password is empty")
|
# raise InvalidPassword("User password is empty")
|
||||||
return {
|
return {"error": "User password is empty"}
|
||||||
"error": "User password is empty"
|
|
||||||
}
|
|
||||||
if not Password.verify(password, user.password):
|
if not Password.verify(password, user.password):
|
||||||
# raise InvalidPassword("Wrong user password")
|
# raise InvalidPassword("Wrong user password")
|
||||||
return {
|
return {"error": "Wrong user password"}
|
||||||
"error": "Wrong user password"
|
|
||||||
}
|
|
||||||
return user
|
return user
|
||||||
|
|
||||||
@staticmethod
|
@staticmethod
|
||||||
|
@ -87,30 +84,22 @@ class Identity:
|
||||||
@staticmethod
|
@staticmethod
|
||||||
async def onetime(token: str) -> User:
|
async def onetime(token: str) -> User:
|
||||||
try:
|
try:
|
||||||
print('[auth.identity] using one time token')
|
print("[auth.identity] using one time token")
|
||||||
payload = JWTCodec.decode(token)
|
payload = JWTCodec.decode(token)
|
||||||
if not await TokenStorage.exist(f"{payload.user_id}-{payload.username}-{token}"):
|
if not await TokenStorage.exist(f"{payload.user_id}-{payload.username}-{token}"):
|
||||||
# raise InvalidToken("Login token has expired, please login again")
|
# raise InvalidToken("Login token has expired, please login again")
|
||||||
return {
|
return {"error": "Token has expired"}
|
||||||
"error": "Token has expired"
|
|
||||||
}
|
|
||||||
except ExpiredSignatureError:
|
except ExpiredSignatureError:
|
||||||
# raise InvalidToken("Login token has expired, please try again")
|
# raise InvalidToken("Login token has expired, please try again")
|
||||||
return {
|
return {"error": "Token has expired"}
|
||||||
"error": "Token has expired"
|
|
||||||
}
|
|
||||||
except DecodeError:
|
except DecodeError:
|
||||||
# raise InvalidToken("token format error") from e
|
# raise InvalidToken("token format error") from e
|
||||||
return {
|
return {"error": "Token format error"}
|
||||||
"error": "Token format error"
|
|
||||||
}
|
|
||||||
with local_session() as session:
|
with local_session() as session:
|
||||||
user = session.query(User).filter_by(id=payload.user_id).first()
|
user = session.query(User).filter_by(id=payload.user_id).first()
|
||||||
if not user:
|
if not user:
|
||||||
# raise Exception("user not exist")
|
# raise Exception("user not exist")
|
||||||
return {
|
return {"error": "User does not exist"}
|
||||||
"error": "User does not exist"
|
|
||||||
}
|
|
||||||
if not user.emailConfirmed:
|
if not user.emailConfirmed:
|
||||||
user.emailConfirmed = True
|
user.emailConfirmed = True
|
||||||
session.commit()
|
session.commit()
|
||||||
|
|
|
@ -1,8 +1,10 @@
|
||||||
from datetime import datetime, timezone
|
from datetime import datetime, timezone
|
||||||
|
|
||||||
import jwt
|
import jwt
|
||||||
|
|
||||||
from base.exceptions import ExpiredToken, InvalidToken
|
from base.exceptions import ExpiredToken, InvalidToken
|
||||||
from validations.auth import TokenPayload, AuthInput
|
|
||||||
from settings import JWT_ALGORITHM, JWT_SECRET_KEY
|
from settings import JWT_ALGORITHM, JWT_SECRET_KEY
|
||||||
|
from validations.auth import AuthInput, TokenPayload
|
||||||
|
|
||||||
|
|
||||||
class JWTCodec:
|
class JWTCodec:
|
||||||
|
@ -13,12 +15,12 @@ class JWTCodec:
|
||||||
"username": user.email or user.phone,
|
"username": user.email or user.phone,
|
||||||
"exp": exp,
|
"exp": exp,
|
||||||
"iat": datetime.now(tz=timezone.utc),
|
"iat": datetime.now(tz=timezone.utc),
|
||||||
"iss": "discours"
|
"iss": "discours",
|
||||||
}
|
}
|
||||||
try:
|
try:
|
||||||
return jwt.encode(payload, JWT_SECRET_KEY, JWT_ALGORITHM)
|
return jwt.encode(payload, JWT_SECRET_KEY, JWT_ALGORITHM)
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
print('[auth.jwtcodec] JWT encode error %r' % e)
|
print("[auth.jwtcodec] JWT encode error %r" % e)
|
||||||
|
|
||||||
@staticmethod
|
@staticmethod
|
||||||
def decode(token: str, verify_exp: bool = True) -> TokenPayload:
|
def decode(token: str, verify_exp: bool = True) -> TokenPayload:
|
||||||
|
@ -33,18 +35,18 @@ class JWTCodec:
|
||||||
# "verify_signature": False
|
# "verify_signature": False
|
||||||
},
|
},
|
||||||
algorithms=[JWT_ALGORITHM],
|
algorithms=[JWT_ALGORITHM],
|
||||||
issuer="discours"
|
issuer="discours",
|
||||||
)
|
)
|
||||||
r = TokenPayload(**payload)
|
r = TokenPayload(**payload)
|
||||||
# print('[auth.jwtcodec] debug token %r' % r)
|
# print('[auth.jwtcodec] debug token %r' % r)
|
||||||
return r
|
return r
|
||||||
except jwt.InvalidIssuedAtError:
|
except jwt.InvalidIssuedAtError:
|
||||||
print('[auth.jwtcodec] invalid issued at: %r' % payload)
|
print("[auth.jwtcodec] invalid issued at: %r" % payload)
|
||||||
raise ExpiredToken('check token issued time')
|
raise ExpiredToken("check token issued time")
|
||||||
except jwt.ExpiredSignatureError:
|
except jwt.ExpiredSignatureError:
|
||||||
print('[auth.jwtcodec] expired signature %r' % payload)
|
print("[auth.jwtcodec] expired signature %r" % payload)
|
||||||
raise ExpiredToken('check token lifetime')
|
raise ExpiredToken("check token lifetime")
|
||||||
except jwt.InvalidTokenError:
|
except jwt.InvalidTokenError:
|
||||||
raise InvalidToken('token is not valid')
|
raise InvalidToken("token is not valid")
|
||||||
except jwt.InvalidSignatureError:
|
except jwt.InvalidSignatureError:
|
||||||
raise InvalidToken('token is not valid')
|
raise InvalidToken("token is not valid")
|
||||||
|
|
|
@ -1,8 +1,9 @@
|
||||||
from authlib.integrations.starlette_client import OAuth
|
from authlib.integrations.starlette_client import OAuth
|
||||||
from starlette.responses import RedirectResponse
|
from starlette.responses import RedirectResponse
|
||||||
|
|
||||||
from auth.identity import Identity
|
from auth.identity import Identity
|
||||||
from auth.tokenstorage import TokenStorage
|
from auth.tokenstorage import TokenStorage
|
||||||
from settings import OAUTH_CLIENTS, FRONTEND_URL
|
from settings import FRONTEND_URL, OAUTH_CLIENTS
|
||||||
|
|
||||||
oauth = OAuth()
|
oauth = OAuth()
|
||||||
|
|
||||||
|
|
|
@ -1,9 +1,9 @@
|
||||||
from datetime import datetime, timedelta, timezone
|
from datetime import datetime, timedelta, timezone
|
||||||
|
|
||||||
from auth.jwtcodec import JWTCodec
|
from auth.jwtcodec import JWTCodec
|
||||||
from validations.auth import AuthInput
|
|
||||||
from base.redis import redis
|
from base.redis import redis
|
||||||
from settings import SESSION_TOKEN_LIFE_SPAN, ONETIME_TOKEN_LIFE_SPAN
|
from settings import ONETIME_TOKEN_LIFE_SPAN, SESSION_TOKEN_LIFE_SPAN
|
||||||
|
from validations.auth import AuthInput
|
||||||
|
|
||||||
|
|
||||||
async def save(token_key, life_span, auto_delete=True):
|
async def save(token_key, life_span, auto_delete=True):
|
||||||
|
@ -35,7 +35,7 @@ class SessionToken:
|
||||||
class TokenStorage:
|
class TokenStorage:
|
||||||
@staticmethod
|
@staticmethod
|
||||||
async def get(token_key):
|
async def get(token_key):
|
||||||
print('[tokenstorage.get] ' + token_key)
|
print("[tokenstorage.get] " + token_key)
|
||||||
# 2041-user@domain.zn-eyJhbGciOiJIUzI1NiIsInR5cCI6IkpXVCJ9.eyJ1c2VyX2lkIjoyMDQxLCJ1c2VybmFtZSI6ImFudG9uLnJld2luK3Rlc3QtbG9hZGNoYXRAZ21haWwuY29tIiwiZXhwIjoxNjcxNzgwNjE2LCJpYXQiOjE2NjkxODg2MTYsImlzcyI6ImRpc2NvdXJzIn0.Nml4oV6iMjMmc6xwM7lTKEZJKBXvJFEIZ-Up1C1rITQ
|
# 2041-user@domain.zn-eyJhbGciOiJIUzI1NiIsInR5cCI6IkpXVCJ9.eyJ1c2VyX2lkIjoyMDQxLCJ1c2VybmFtZSI6ImFudG9uLnJld2luK3Rlc3QtbG9hZGNoYXRAZ21haWwuY29tIiwiZXhwIjoxNjcxNzgwNjE2LCJpYXQiOjE2NjkxODg2MTYsImlzcyI6ImRpc2NvdXJzIn0.Nml4oV6iMjMmc6xwM7lTKEZJKBXvJFEIZ-Up1C1rITQ
|
||||||
return await redis.execute("GET", token_key)
|
return await redis.execute("GET", token_key)
|
||||||
|
|
||||||
|
|
|
@ -1,8 +1,8 @@
|
||||||
from graphql.error import GraphQLError
|
from graphql.error import GraphQLError
|
||||||
|
|
||||||
|
|
||||||
# TODO: remove traceback from logs for defined exceptions
|
# TODO: remove traceback from logs for defined exceptions
|
||||||
|
|
||||||
|
|
||||||
class BaseHttpException(GraphQLError):
|
class BaseHttpException(GraphQLError):
|
||||||
code = 500
|
code = 500
|
||||||
message = "500 Server error"
|
message = "500 Server error"
|
||||||
|
|
17
base/orm.py
17
base/orm.py
|
@ -1,15 +1,13 @@
|
||||||
from typing import TypeVar, Any, Dict, Generic, Callable
|
from typing import Any, Callable, Dict, Generic, TypeVar
|
||||||
|
|
||||||
from sqlalchemy import create_engine, Column, Integer
|
from sqlalchemy import Column, Integer, create_engine
|
||||||
from sqlalchemy.ext.declarative import declarative_base
|
from sqlalchemy.ext.declarative import declarative_base
|
||||||
from sqlalchemy.orm import Session
|
from sqlalchemy.orm import Session
|
||||||
from sqlalchemy.sql.schema import Table
|
from sqlalchemy.sql.schema import Table
|
||||||
|
|
||||||
from settings import DB_URL
|
from settings import DB_URL
|
||||||
|
|
||||||
engine = create_engine(
|
engine = create_engine(DB_URL, echo=False, pool_size=10, max_overflow=20)
|
||||||
DB_URL, echo=False, pool_size=10, max_overflow=20
|
|
||||||
)
|
|
||||||
|
|
||||||
T = TypeVar("T")
|
T = TypeVar("T")
|
||||||
|
|
||||||
|
@ -20,7 +18,10 @@ def local_session():
|
||||||
return Session(bind=engine, expire_on_commit=False)
|
return Session(bind=engine, expire_on_commit=False)
|
||||||
|
|
||||||
|
|
||||||
class Base(declarative_base()):
|
DeclarativeBase = declarative_base() # type: Any
|
||||||
|
|
||||||
|
|
||||||
|
class Base(DeclarativeBase):
|
||||||
__table__: Table
|
__table__: Table
|
||||||
__tablename__: str
|
__tablename__: str
|
||||||
__new__: Callable
|
__new__: Callable
|
||||||
|
@ -29,7 +30,7 @@ class Base(declarative_base()):
|
||||||
__abstract__ = True
|
__abstract__ = True
|
||||||
__table_args__ = {"extend_existing": True}
|
__table_args__ = {"extend_existing": True}
|
||||||
|
|
||||||
id = Column(Integer, primary_key=True)
|
id: Column | None = Column(Integer, primary_key=True)
|
||||||
|
|
||||||
def __init_subclass__(cls, **kwargs):
|
def __init_subclass__(cls, **kwargs):
|
||||||
REGISTRY[cls.__name__] = cls
|
REGISTRY[cls.__name__] = cls
|
||||||
|
@ -47,7 +48,7 @@ class Base(declarative_base()):
|
||||||
|
|
||||||
def update(self, input):
|
def update(self, input):
|
||||||
column_names = self.__table__.columns.keys()
|
column_names = self.__table__.columns.keys()
|
||||||
for (name, value) in input.items():
|
for name, value in input.items():
|
||||||
if name in column_names:
|
if name in column_names:
|
||||||
setattr(self, name, value)
|
setattr(self, name, value)
|
||||||
|
|
||||||
|
|
|
@ -1,5 +1,7 @@
|
||||||
from aioredis import from_url
|
|
||||||
from asyncio import sleep
|
from asyncio import sleep
|
||||||
|
|
||||||
|
from aioredis import from_url
|
||||||
|
|
||||||
from settings import REDIS_URL
|
from settings import REDIS_URL
|
||||||
|
|
||||||
|
|
||||||
|
|
1
generate_gql_types.sh
Executable file
1
generate_gql_types.sh
Executable file
|
@ -0,0 +1 @@
|
||||||
|
python -m gql_schema_codegen -p ./schema.graphql -t ./schema_types.py
|
18
lint.sh
18
lint.sh
|
@ -1,16 +1,10 @@
|
||||||
#!/usr/bin/env bash
|
#!/usr/bin/env bash
|
||||||
set -e
|
|
||||||
|
|
||||||
find . -name "*.py[co]" -o -name __pycache__ -exec rm -rf {} +
|
|
||||||
#rm -rf .mypy_cache
|
|
||||||
|
|
||||||
echo "> isort"
|
echo "> isort"
|
||||||
isort --gitignore --settings-file=setup.cfg .
|
isort .
|
||||||
echo "> brunette"
|
echo "> black"
|
||||||
brunette --config=setup.cfg .
|
black .
|
||||||
echo "> flake8"
|
echo "> flake8"
|
||||||
flake8 --config=setup.cfg .
|
flake8 .
|
||||||
echo "> mypy"
|
# echo "> mypy"
|
||||||
mypy --config-file=setup.cfg .
|
# mypy .
|
||||||
echo "> prettyjson"
|
|
||||||
python3 -m scripts.prettyjson
|
|
||||||
|
|
26
main.py
26
main.py
|
@ -2,6 +2,7 @@ import asyncio
|
||||||
import os
|
import os
|
||||||
from importlib import import_module
|
from importlib import import_module
|
||||||
from os.path import exists
|
from os.path import exists
|
||||||
|
|
||||||
from ariadne import load_schema_from_path, make_executable_schema
|
from ariadne import load_schema_from_path, make_executable_schema
|
||||||
from ariadne.asgi import GraphQL
|
from ariadne.asgi import GraphQL
|
||||||
from starlette.applications import Starlette
|
from starlette.applications import Starlette
|
||||||
|
@ -9,23 +10,24 @@ from starlette.middleware import Middleware
|
||||||
from starlette.middleware.authentication import AuthenticationMiddleware
|
from starlette.middleware.authentication import AuthenticationMiddleware
|
||||||
from starlette.middleware.sessions import SessionMiddleware
|
from starlette.middleware.sessions import SessionMiddleware
|
||||||
from starlette.routing import Route
|
from starlette.routing import Route
|
||||||
from orm import init_tables
|
|
||||||
|
|
||||||
from auth.authenticate import JWTAuthenticate
|
from auth.authenticate import JWTAuthenticate
|
||||||
from auth.oauth import oauth_login, oauth_authorize
|
from auth.oauth import oauth_authorize, oauth_login
|
||||||
from base.redis import redis
|
from base.redis import redis
|
||||||
from base.resolvers import resolvers
|
from base.resolvers import resolvers
|
||||||
|
from orm import init_tables
|
||||||
from resolvers.auth import confirm_email_handler
|
from resolvers.auth import confirm_email_handler
|
||||||
from resolvers.upload import upload_handler
|
from resolvers.upload import upload_handler
|
||||||
from services.main import storages_init
|
from services.main import storages_init
|
||||||
from services.notifications.notification_service import notification_service
|
from services.notifications.notification_service import notification_service
|
||||||
|
from services.notifications.sse import sse_subscribe_handler
|
||||||
from services.stat.viewed import ViewedStorage
|
from services.stat.viewed import ViewedStorage
|
||||||
|
|
||||||
# from services.zine.gittask import GitTask
|
# from services.zine.gittask import GitTask
|
||||||
from settings import DEV_SERVER_PID_FILE_NAME, SENTRY_DSN, SESSION_SECRET_KEY
|
from settings import DEV_SERVER_PID_FILE_NAME, SENTRY_DSN, SESSION_SECRET_KEY
|
||||||
from services.notifications.sse import sse_subscribe_handler
|
|
||||||
|
|
||||||
import_module("resolvers")
|
import_module("resolvers")
|
||||||
schema = make_executable_schema(load_schema_from_path("schema.graphql"), resolvers) # type: ignore
|
schema = make_executable_schema(load_schema_from_path("schema.graphql"), resolvers)
|
||||||
|
|
||||||
middleware = [
|
middleware = [
|
||||||
Middleware(AuthenticationMiddleware, backend=JWTAuthenticate()),
|
Middleware(AuthenticationMiddleware, backend=JWTAuthenticate()),
|
||||||
|
@ -46,9 +48,10 @@ async def start_up():
|
||||||
|
|
||||||
try:
|
try:
|
||||||
import sentry_sdk
|
import sentry_sdk
|
||||||
|
|
||||||
sentry_sdk.init(SENTRY_DSN)
|
sentry_sdk.init(SENTRY_DSN)
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
print('[sentry] init error')
|
print("[sentry] init error")
|
||||||
print(e)
|
print(e)
|
||||||
|
|
||||||
|
|
||||||
|
@ -57,7 +60,7 @@ async def dev_start_up():
|
||||||
await redis.connect()
|
await redis.connect()
|
||||||
return
|
return
|
||||||
else:
|
else:
|
||||||
with open(DEV_SERVER_PID_FILE_NAME, 'w', encoding='utf-8') as f:
|
with open(DEV_SERVER_PID_FILE_NAME, "w", encoding="utf-8") as f:
|
||||||
f.write(str(os.getpid()))
|
f.write(str(os.getpid()))
|
||||||
|
|
||||||
await start_up()
|
await start_up()
|
||||||
|
@ -72,7 +75,7 @@ routes = [
|
||||||
Route("/oauth/{provider}", endpoint=oauth_login),
|
Route("/oauth/{provider}", endpoint=oauth_login),
|
||||||
Route("/oauth-authorize", endpoint=oauth_authorize),
|
Route("/oauth-authorize", endpoint=oauth_authorize),
|
||||||
Route("/confirm/{token}", endpoint=confirm_email_handler),
|
Route("/confirm/{token}", endpoint=confirm_email_handler),
|
||||||
Route("/upload", endpoint=upload_handler, methods=['POST']),
|
Route("/upload", endpoint=upload_handler, methods=["POST"]),
|
||||||
Route("/subscribe/{user_id}", endpoint=sse_subscribe_handler),
|
Route("/subscribe/{user_id}", endpoint=sse_subscribe_handler),
|
||||||
]
|
]
|
||||||
|
|
||||||
|
@ -82,9 +85,7 @@ app = Starlette(
|
||||||
middleware=middleware,
|
middleware=middleware,
|
||||||
routes=routes,
|
routes=routes,
|
||||||
)
|
)
|
||||||
app.mount("/", GraphQL(
|
app.mount("/", GraphQL(schema))
|
||||||
schema
|
|
||||||
))
|
|
||||||
|
|
||||||
dev_app = Starlette(
|
dev_app = Starlette(
|
||||||
debug=True,
|
debug=True,
|
||||||
|
@ -93,7 +94,4 @@ dev_app = Starlette(
|
||||||
middleware=middleware,
|
middleware=middleware,
|
||||||
routes=routes,
|
routes=routes,
|
||||||
)
|
)
|
||||||
dev_app.mount("/", GraphQL(
|
dev_app.mount("/", GraphQL(schema, debug=True))
|
||||||
schema,
|
|
||||||
debug=True
|
|
||||||
))
|
|
||||||
|
|
|
@ -16,4 +16,3 @@ echo "Start migration"
|
||||||
python3 server.py migrate
|
python3 server.py migrate
|
||||||
if [ $? -ne 0 ]; then { echo "Migration failed, aborting." ; exit 1; } fi
|
if [ $? -ne 0 ]; then { echo "Migration failed, aborting." ; exit 1; } fi
|
||||||
echo 'Done!'
|
echo 'Done!'
|
||||||
|
|
||||||
|
|
|
@ -12,10 +12,12 @@ from migration.tables.comments import migrate as migrateComment
|
||||||
from migration.tables.comments import migrate_2stage as migrateComment_2stage
|
from migration.tables.comments import migrate_2stage as migrateComment_2stage
|
||||||
from migration.tables.content_items import get_shout_slug
|
from migration.tables.content_items import get_shout_slug
|
||||||
from migration.tables.content_items import migrate as migrateShout
|
from migration.tables.content_items import migrate as migrateShout
|
||||||
from migration.tables.remarks import migrate as migrateRemark
|
|
||||||
|
# from migration.tables.remarks import migrate as migrateRemark
|
||||||
from migration.tables.topics import migrate as migrateTopic
|
from migration.tables.topics import migrate as migrateTopic
|
||||||
from migration.tables.users import migrate as migrateUser, post_migrate as users_post_migrate
|
from migration.tables.users import migrate as migrateUser
|
||||||
from migration.tables.users import migrate_2stage as migrateUser_2stage
|
from migration.tables.users import migrate_2stage as migrateUser_2stage
|
||||||
|
from migration.tables.users import post_migrate as users_post_migrate
|
||||||
from orm import init_tables
|
from orm import init_tables
|
||||||
from orm.reaction import Reaction
|
from orm.reaction import Reaction
|
||||||
|
|
||||||
|
@ -63,16 +65,8 @@ async def topics_handle(storage):
|
||||||
del storage["topics"]["by_slug"][oldslug]
|
del storage["topics"]["by_slug"][oldslug]
|
||||||
storage["topics"]["by_oid"][oid] = storage["topics"]["by_slug"][newslug]
|
storage["topics"]["by_oid"][oid] = storage["topics"]["by_slug"][newslug]
|
||||||
print("[migration] " + str(counter) + " topics migrated")
|
print("[migration] " + str(counter) + " topics migrated")
|
||||||
print(
|
print("[migration] " + str(len(storage["topics"]["by_oid"].values())) + " topics by oid")
|
||||||
"[migration] "
|
print("[migration] " + str(len(storage["topics"]["by_slug"].values())) + " topics by slug")
|
||||||
+ str(len(storage["topics"]["by_oid"].values()))
|
|
||||||
+ " topics by oid"
|
|
||||||
)
|
|
||||||
print(
|
|
||||||
"[migration] "
|
|
||||||
+ str(len(storage["topics"]["by_slug"].values()))
|
|
||||||
+ " topics by slug"
|
|
||||||
)
|
|
||||||
|
|
||||||
|
|
||||||
async def shouts_handle(storage, args):
|
async def shouts_handle(storage, args):
|
||||||
|
@ -117,9 +111,10 @@ async def shouts_handle(storage, args):
|
||||||
|
|
||||||
# print main counter
|
# print main counter
|
||||||
counter += 1
|
counter += 1
|
||||||
print('[migration] shouts_handle %d: %s @%s' % (
|
print(
|
||||||
(counter + 1), shout_dict["slug"], author["slug"]
|
"[migration] shouts_handle %d: %s @%s"
|
||||||
))
|
% ((counter + 1), shout_dict["slug"], author["slug"])
|
||||||
|
)
|
||||||
|
|
||||||
b = bs4.BeautifulSoup(shout_dict["body"], "html.parser")
|
b = bs4.BeautifulSoup(shout_dict["body"], "html.parser")
|
||||||
texts = [shout_dict["title"].lower().replace(r"[^а-яА-Яa-zA-Z]", "")]
|
texts = [shout_dict["title"].lower().replace(r"[^а-яА-Яa-zA-Z]", "")]
|
||||||
|
@ -138,13 +133,13 @@ async def shouts_handle(storage, args):
|
||||||
print("[migration] " + str(anonymous_author) + " authored by @anonymous")
|
print("[migration] " + str(anonymous_author) + " authored by @anonymous")
|
||||||
|
|
||||||
|
|
||||||
async def remarks_handle(storage):
|
# async def remarks_handle(storage):
|
||||||
print("[migration] comments")
|
# print("[migration] comments")
|
||||||
c = 0
|
# c = 0
|
||||||
for entry_remark in storage["remarks"]["data"]:
|
# for entry_remark in storage["remarks"]["data"]:
|
||||||
remark = await migrateRemark(entry_remark, storage)
|
# remark = await migrateRemark(entry_remark, storage)
|
||||||
c += 1
|
# c += 1
|
||||||
print("[migration] " + str(c) + " remarks migrated")
|
# print("[migration] " + str(c) + " remarks migrated")
|
||||||
|
|
||||||
|
|
||||||
async def comments_handle(storage):
|
async def comments_handle(storage):
|
||||||
|
@ -155,9 +150,9 @@ async def comments_handle(storage):
|
||||||
for oldcomment in storage["reactions"]["data"]:
|
for oldcomment in storage["reactions"]["data"]:
|
||||||
if not oldcomment.get("deleted"):
|
if not oldcomment.get("deleted"):
|
||||||
reaction = await migrateComment(oldcomment, storage)
|
reaction = await migrateComment(oldcomment, storage)
|
||||||
if type(reaction) == str:
|
if isinstance(reaction, str):
|
||||||
missed_shouts[reaction] = oldcomment
|
missed_shouts[reaction] = oldcomment
|
||||||
elif type(reaction) == Reaction:
|
elif isinstance(reaction, Reaction):
|
||||||
reaction = reaction.dict()
|
reaction = reaction.dict()
|
||||||
rid = reaction["id"]
|
rid = reaction["id"]
|
||||||
oid = reaction["oid"]
|
oid = reaction["oid"]
|
||||||
|
@ -214,9 +209,7 @@ def data_load():
|
||||||
tags_data = json.loads(open("migration/data/tags.json").read())
|
tags_data = json.loads(open("migration/data/tags.json").read())
|
||||||
storage["topics"]["tags"] = tags_data
|
storage["topics"]["tags"] = tags_data
|
||||||
print("[migration.load] " + str(len(tags_data)) + " tags ")
|
print("[migration.load] " + str(len(tags_data)) + " tags ")
|
||||||
cats_data = json.loads(
|
cats_data = json.loads(open("migration/data/content_item_categories.json").read())
|
||||||
open("migration/data/content_item_categories.json").read()
|
|
||||||
)
|
|
||||||
storage["topics"]["cats"] = cats_data
|
storage["topics"]["cats"] = cats_data
|
||||||
print("[migration.load] " + str(len(cats_data)) + " cats ")
|
print("[migration.load] " + str(len(cats_data)) + " cats ")
|
||||||
comments_data = json.loads(open("migration/data/comments.json").read())
|
comments_data = json.loads(open("migration/data/comments.json").read())
|
||||||
|
@ -235,11 +228,7 @@ def data_load():
|
||||||
storage["users"]["by_oid"][x["_id"]] = x
|
storage["users"]["by_oid"][x["_id"]] = x
|
||||||
# storage['users']['by_slug'][x['slug']] = x
|
# storage['users']['by_slug'][x['slug']] = x
|
||||||
# no user.slug yet
|
# no user.slug yet
|
||||||
print(
|
print("[migration.load] " + str(len(storage["users"]["by_oid"].keys())) + " users by oid")
|
||||||
"[migration.load] "
|
|
||||||
+ str(len(storage["users"]["by_oid"].keys()))
|
|
||||||
+ " users by oid"
|
|
||||||
)
|
|
||||||
for x in tags_data:
|
for x in tags_data:
|
||||||
storage["topics"]["by_oid"][x["_id"]] = x
|
storage["topics"]["by_oid"][x["_id"]] = x
|
||||||
storage["topics"]["by_slug"][x["slug"]] = x
|
storage["topics"]["by_slug"][x["slug"]] = x
|
||||||
|
@ -247,9 +236,7 @@ def data_load():
|
||||||
storage["topics"]["by_oid"][x["_id"]] = x
|
storage["topics"]["by_oid"][x["_id"]] = x
|
||||||
storage["topics"]["by_slug"][x["slug"]] = x
|
storage["topics"]["by_slug"][x["slug"]] = x
|
||||||
print(
|
print(
|
||||||
"[migration.load] "
|
"[migration.load] " + str(len(storage["topics"]["by_slug"].keys())) + " topics by slug"
|
||||||
+ str(len(storage["topics"]["by_slug"].keys()))
|
|
||||||
+ " topics by slug"
|
|
||||||
)
|
)
|
||||||
for item in content_data:
|
for item in content_data:
|
||||||
slug = get_shout_slug(item)
|
slug = get_shout_slug(item)
|
||||||
|
|
|
@ -1,8 +1,9 @@
|
||||||
|
import gc
|
||||||
import json
|
import json
|
||||||
import os
|
import os
|
||||||
|
|
||||||
import bson
|
import bson
|
||||||
import gc
|
|
||||||
from .utils import DateTimeEncoder
|
from .utils import DateTimeEncoder
|
||||||
|
|
||||||
|
|
||||||
|
@ -15,10 +16,10 @@ def json_tables():
|
||||||
"email_subscriptions": [],
|
"email_subscriptions": [],
|
||||||
"users": [],
|
"users": [],
|
||||||
"comments": [],
|
"comments": [],
|
||||||
"remarks": []
|
"remarks": [],
|
||||||
}
|
}
|
||||||
for table in data.keys():
|
for table in data.keys():
|
||||||
print('[migration] bson2json for ' + table)
|
print("[migration] bson2json for " + table)
|
||||||
gc.collect()
|
gc.collect()
|
||||||
lc = []
|
lc = []
|
||||||
bs = open("dump/discours/" + table + ".bson", "rb").read()
|
bs = open("dump/discours/" + table + ".bson", "rb").read()
|
||||||
|
|
|
@ -71,47 +71,29 @@ def export_slug(slug, storage):
|
||||||
|
|
||||||
|
|
||||||
def export_email_subscriptions():
|
def export_email_subscriptions():
|
||||||
email_subscriptions_data = json.loads(
|
email_subscriptions_data = json.loads(open("migration/data/email_subscriptions.json").read())
|
||||||
open("migration/data/email_subscriptions.json").read()
|
|
||||||
)
|
|
||||||
for data in email_subscriptions_data:
|
for data in email_subscriptions_data:
|
||||||
# TODO: migrate to mailgun list manually
|
# TODO: migrate to mailgun list manually
|
||||||
# migrate_email_subscription(data)
|
# migrate_email_subscription(data)
|
||||||
pass
|
pass
|
||||||
print(
|
print("[migration] " + str(len(email_subscriptions_data)) + " email subscriptions exported")
|
||||||
"[migration] "
|
|
||||||
+ str(len(email_subscriptions_data))
|
|
||||||
+ " email subscriptions exported"
|
|
||||||
)
|
|
||||||
|
|
||||||
|
|
||||||
def export_shouts(storage):
|
def export_shouts(storage):
|
||||||
# update what was just migrated or load json again
|
# update what was just migrated or load json again
|
||||||
if len(storage["users"]["by_slugs"].keys()) == 0:
|
if len(storage["users"]["by_slugs"].keys()) == 0:
|
||||||
storage["users"]["by_slugs"] = json.loads(
|
storage["users"]["by_slugs"] = json.loads(open(EXPORT_DEST + "authors.json").read())
|
||||||
open(EXPORT_DEST + "authors.json").read()
|
print("[migration] " + str(len(storage["users"]["by_slugs"].keys())) + " exported authors ")
|
||||||
)
|
|
||||||
print(
|
|
||||||
"[migration] "
|
|
||||||
+ str(len(storage["users"]["by_slugs"].keys()))
|
|
||||||
+ " exported authors "
|
|
||||||
)
|
|
||||||
if len(storage["shouts"]["by_slugs"].keys()) == 0:
|
if len(storage["shouts"]["by_slugs"].keys()) == 0:
|
||||||
storage["shouts"]["by_slugs"] = json.loads(
|
storage["shouts"]["by_slugs"] = json.loads(open(EXPORT_DEST + "articles.json").read())
|
||||||
open(EXPORT_DEST + "articles.json").read()
|
|
||||||
)
|
|
||||||
print(
|
print(
|
||||||
"[migration] "
|
"[migration] " + str(len(storage["shouts"]["by_slugs"].keys())) + " exported articles "
|
||||||
+ str(len(storage["shouts"]["by_slugs"].keys()))
|
|
||||||
+ " exported articles "
|
|
||||||
)
|
)
|
||||||
for slug in storage["shouts"]["by_slugs"].keys():
|
for slug in storage["shouts"]["by_slugs"].keys():
|
||||||
export_slug(slug, storage)
|
export_slug(slug, storage)
|
||||||
|
|
||||||
|
|
||||||
def export_json(
|
def export_json(export_articles={}, export_authors={}, export_topics={}, export_comments={}):
|
||||||
export_articles={}, export_authors={}, export_topics={}, export_comments={}
|
|
||||||
):
|
|
||||||
open(EXPORT_DEST + "authors.json", "w").write(
|
open(EXPORT_DEST + "authors.json", "w").write(
|
||||||
json.dumps(
|
json.dumps(
|
||||||
export_authors,
|
export_authors,
|
||||||
|
@ -152,8 +134,4 @@ def export_json(
|
||||||
ensure_ascii=False,
|
ensure_ascii=False,
|
||||||
)
|
)
|
||||||
)
|
)
|
||||||
print(
|
print("[migration] " + str(len(export_comments.items())) + " exported articles with comments")
|
||||||
"[migration] "
|
|
||||||
+ str(len(export_comments.items()))
|
|
||||||
+ " exported articles with comments"
|
|
||||||
)
|
|
||||||
|
|
|
@ -1,11 +1,8 @@
|
||||||
import base64
|
|
||||||
import os
|
import os
|
||||||
import re
|
import re
|
||||||
import uuid
|
|
||||||
|
|
||||||
from bs4 import BeautifulSoup
|
from bs4 import BeautifulSoup
|
||||||
|
|
||||||
|
|
||||||
TOOLTIP_REGEX = r"(\/\/\/(.+)\/\/\/)"
|
TOOLTIP_REGEX = r"(\/\/\/(.+)\/\/\/)"
|
||||||
contentDir = os.path.join(
|
contentDir = os.path.join(
|
||||||
os.path.dirname(os.path.realpath(__file__)), "..", "..", "discoursio-web", "content"
|
os.path.dirname(os.path.realpath(__file__)), "..", "..", "discoursio-web", "content"
|
||||||
|
@ -27,76 +24,79 @@ def replace_tooltips(body):
|
||||||
return newbody
|
return newbody
|
||||||
|
|
||||||
|
|
||||||
|
# def extract_footnotes(body, shout_dict):
|
||||||
def extract_footnotes(body, shout_dict):
|
# parts = body.split("&&&")
|
||||||
parts = body.split("&&&")
|
# lll = len(parts)
|
||||||
lll = len(parts)
|
# newparts = list(parts)
|
||||||
newparts = list(parts)
|
# placed = False
|
||||||
placed = False
|
# if lll & 1:
|
||||||
if lll & 1:
|
# if lll > 1:
|
||||||
if lll > 1:
|
# i = 1
|
||||||
i = 1
|
# print("[extract] found %d footnotes in body" % (lll - 1))
|
||||||
print("[extract] found %d footnotes in body" % (lll - 1))
|
# for part in parts[1:]:
|
||||||
for part in parts[1:]:
|
# if i & 1:
|
||||||
if i & 1:
|
# placed = True
|
||||||
placed = True
|
# if 'a class="footnote-url" href=' in part:
|
||||||
if 'a class="footnote-url" href=' in part:
|
# print("[extract] footnote: " + part)
|
||||||
print("[extract] footnote: " + part)
|
# fn = 'a class="footnote-url" href="'
|
||||||
fn = 'a class="footnote-url" href="'
|
# exxtracted_link = part.split(fn, 1)[1].split('"', 1)[0]
|
||||||
exxtracted_link = part.split(fn, 1)[1].split('"', 1)[0]
|
# extracted_body = part.split(fn, 1)[1].split(">", 1)[1].split("</a>", 1)[0]
|
||||||
extracted_body = part.split(fn, 1)[1].split('>', 1)[1].split('</a>', 1)[0]
|
# print("[extract] footnote link: " + extracted_link)
|
||||||
print("[extract] footnote link: " + extracted_link)
|
# with local_session() as session:
|
||||||
with local_session() as session:
|
# Reaction.create(
|
||||||
Reaction.create({
|
# {
|
||||||
"shout": shout_dict['id'],
|
# "shout": shout_dict["id"],
|
||||||
"kind": ReactionKind.FOOTNOTE,
|
# "kind": ReactionKind.FOOTNOTE,
|
||||||
"body": extracted_body,
|
# "body": extracted_body,
|
||||||
"range": str(body.index(fn + link) - len('<')) + ':' + str(body.index(extracted_body) + len('</a>'))
|
# "range": str(body.index(fn + link) - len("<"))
|
||||||
})
|
# + ":"
|
||||||
newparts[i] = "<a href='#'>ℹ️</a>"
|
# + str(body.index(extracted_body) + len("</a>")),
|
||||||
else:
|
# }
|
||||||
newparts[i] = part
|
# )
|
||||||
i += 1
|
# newparts[i] = "<a href='#'>ℹ️</a>"
|
||||||
return ("".join(newparts), placed)
|
# else:
|
||||||
|
# newparts[i] = part
|
||||||
|
# i += 1
|
||||||
|
# return ("".join(newparts), placed)
|
||||||
|
|
||||||
|
|
||||||
def place_tooltips(body):
|
# def place_tooltips(body):
|
||||||
parts = body.split("&&&")
|
# parts = body.split("&&&")
|
||||||
lll = len(parts)
|
# lll = len(parts)
|
||||||
newparts = list(parts)
|
# newparts = list(parts)
|
||||||
placed = False
|
# placed = False
|
||||||
if lll & 1:
|
# if lll & 1:
|
||||||
if lll > 1:
|
# if lll > 1:
|
||||||
i = 1
|
# i = 1
|
||||||
print("[extract] found %d tooltips" % (lll - 1))
|
# print("[extract] found %d tooltips" % (lll - 1))
|
||||||
for part in parts[1:]:
|
# for part in parts[1:]:
|
||||||
if i & 1:
|
# if i & 1:
|
||||||
placed = True
|
# placed = True
|
||||||
if 'a class="footnote-url" href=' in part:
|
# if 'a class="footnote-url" href=' in part:
|
||||||
print("[extract] footnote: " + part)
|
# print("[extract] footnote: " + part)
|
||||||
fn = 'a class="footnote-url" href="'
|
# fn = 'a class="footnote-url" href="'
|
||||||
link = part.split(fn, 1)[1].split('"', 1)[0]
|
# link = part.split(fn, 1)[1].split('"', 1)[0]
|
||||||
extracted_part = (
|
# extracted_part = part.split(fn, 1)[0] + " " + part.split("/", 1)[-1]
|
||||||
part.split(fn, 1)[0] + " " + part.split("/", 1)[-1]
|
# newparts[i] = (
|
||||||
)
|
# "<Tooltip"
|
||||||
newparts[i] = (
|
# + (' link="' + link + '" ' if link else "")
|
||||||
"<Tooltip"
|
# + ">"
|
||||||
+ (' link="' + link + '" ' if link else "")
|
# + extracted_part
|
||||||
+ ">"
|
# + "</Tooltip>"
|
||||||
+ extracted_part
|
# )
|
||||||
+ "</Tooltip>"
|
# else:
|
||||||
)
|
# newparts[i] = "<Tooltip>%s</Tooltip>" % part
|
||||||
else:
|
# # print('[extract] ' + newparts[i])
|
||||||
newparts[i] = "<Tooltip>%s</Tooltip>" % part
|
# else:
|
||||||
# print('[extract] ' + newparts[i])
|
# # print('[extract] ' + part[:10] + '..')
|
||||||
else:
|
# newparts[i] = part
|
||||||
# print('[extract] ' + part[:10] + '..')
|
# i += 1
|
||||||
newparts[i] = part
|
# return ("".join(newparts), placed)
|
||||||
i += 1
|
|
||||||
return ("".join(newparts), placed)
|
|
||||||
|
|
||||||
|
|
||||||
IMG_REGEX = r"\!\[(.*?)\]\((data\:image\/(png|jpeg|jpg);base64\,((?:[A-Za-z\d+\/]{4})*(?:[A-Za-z\d+\/]{3}="
|
IMG_REGEX = (
|
||||||
|
r"\!\[(.*?)\]\((data\:image\/(png|jpeg|jpg);base64\,((?:[A-Za-z\d+\/]{4})*(?:[A-Za-z\d+\/]{3}="
|
||||||
|
)
|
||||||
IMG_REGEX += r"|[A-Za-z\d+\/]{2}==)))\)"
|
IMG_REGEX += r"|[A-Za-z\d+\/]{2}==)))\)"
|
||||||
|
|
||||||
parentDir = "/".join(os.getcwd().split("/")[:-1])
|
parentDir = "/".join(os.getcwd().split("/")[:-1])
|
||||||
|
@ -104,29 +104,29 @@ public = parentDir + "/discoursio-web/public"
|
||||||
cache = {}
|
cache = {}
|
||||||
|
|
||||||
|
|
||||||
def reextract_images(body, oid):
|
# def reextract_images(body, oid):
|
||||||
# change if you prefer regexp
|
# # change if you prefer regexp
|
||||||
matches = list(re.finditer(IMG_REGEX, body, re.IGNORECASE | re.MULTILINE))[1:]
|
# matches = list(re.finditer(IMG_REGEX, body, re.IGNORECASE | re.MULTILINE))[1:]
|
||||||
i = 0
|
# i = 0
|
||||||
for match in matches:
|
# for match in matches:
|
||||||
print("[extract] image " + match.group(1))
|
# print("[extract] image " + match.group(1))
|
||||||
ext = match.group(3)
|
# ext = match.group(3)
|
||||||
name = oid + str(i)
|
# name = oid + str(i)
|
||||||
link = public + "/upload/image-" + name + "." + ext
|
# link = public + "/upload/image-" + name + "." + ext
|
||||||
img = match.group(4)
|
# img = match.group(4)
|
||||||
title = match.group(1) # NOTE: this is not the title
|
# title = match.group(1) # NOTE: this is not the title
|
||||||
if img not in cache:
|
# if img not in cache:
|
||||||
content = base64.b64decode(img + "==")
|
# content = base64.b64decode(img + "==")
|
||||||
print(str(len(img)) + " image bytes been written")
|
# print(str(len(img)) + " image bytes been written")
|
||||||
open("../" + link, "wb").write(content)
|
# open("../" + link, "wb").write(content)
|
||||||
cache[img] = name
|
# cache[img] = name
|
||||||
i += 1
|
# i += 1
|
||||||
else:
|
# else:
|
||||||
print("[extract] image cached " + cache[img])
|
# print("[extract] image cached " + cache[img])
|
||||||
body.replace(
|
# body.replace(
|
||||||
str(match), ""
|
# str(match), ""
|
||||||
) # WARNING: this does not work
|
# ) # WARNING: this does not work
|
||||||
return body
|
# return body
|
||||||
|
|
||||||
|
|
||||||
IMAGES = {
|
IMAGES = {
|
||||||
|
@ -137,163 +137,11 @@ IMAGES = {
|
||||||
|
|
||||||
b64 = ";base64,"
|
b64 = ";base64,"
|
||||||
|
|
||||||
|
|
||||||
def extract_imageparts(bodyparts, prefix):
|
|
||||||
# recursive loop
|
|
||||||
newparts = list(bodyparts)
|
|
||||||
for current in bodyparts:
|
|
||||||
i = bodyparts.index(current)
|
|
||||||
for mime in IMAGES.keys():
|
|
||||||
if mime == current[-len(mime) :] and (i + 1 < len(bodyparts)):
|
|
||||||
print("[extract] " + mime)
|
|
||||||
next = bodyparts[i + 1]
|
|
||||||
ext = IMAGES[mime]
|
|
||||||
b64end = next.index(")")
|
|
||||||
b64encoded = next[:b64end]
|
|
||||||
name = prefix + "-" + str(len(cache))
|
|
||||||
link = "/upload/image-" + name + "." + ext
|
|
||||||
print("[extract] name: " + name)
|
|
||||||
print("[extract] link: " + link)
|
|
||||||
print("[extract] %d bytes" % len(b64encoded))
|
|
||||||
if b64encoded not in cache:
|
|
||||||
try:
|
|
||||||
content = base64.b64decode(b64encoded + "==")
|
|
||||||
open(public + link, "wb").write(content)
|
|
||||||
print(
|
|
||||||
"[extract] "
|
|
||||||
+ str(len(content))
|
|
||||||
+ " image bytes been written"
|
|
||||||
)
|
|
||||||
cache[b64encoded] = name
|
|
||||||
except Exception:
|
|
||||||
raise Exception
|
|
||||||
# raise Exception('[extract] error decoding image %r' %b64encoded)
|
|
||||||
else:
|
|
||||||
print("[extract] cached link " + cache[b64encoded])
|
|
||||||
name = cache[b64encoded]
|
|
||||||
link = cdn + "/upload/image-" + name + "." + ext
|
|
||||||
newparts[i] = (
|
|
||||||
current[: -len(mime)]
|
|
||||||
+ current[-len(mime) :]
|
|
||||||
+ link
|
|
||||||
+ next[-b64end:]
|
|
||||||
)
|
|
||||||
newparts[i + 1] = next[:-b64end]
|
|
||||||
break
|
|
||||||
return (
|
|
||||||
extract_imageparts(
|
|
||||||
newparts[i] + newparts[i + 1] + b64.join(bodyparts[(i + 2) :]), prefix
|
|
||||||
)
|
|
||||||
if len(bodyparts) > (i + 1)
|
|
||||||
else "".join(newparts)
|
|
||||||
)
|
|
||||||
|
|
||||||
|
|
||||||
def extract_dataimages(parts, prefix):
|
|
||||||
newparts = list(parts)
|
|
||||||
for part in parts:
|
|
||||||
i = parts.index(part)
|
|
||||||
if part.endswith("]("):
|
|
||||||
[ext, rest] = parts[i + 1].split(b64)
|
|
||||||
name = prefix + "-" + str(len(cache))
|
|
||||||
if ext == "/jpeg":
|
|
||||||
ext = "jpg"
|
|
||||||
else:
|
|
||||||
ext = ext.replace("/", "")
|
|
||||||
link = "/upload/image-" + name + "." + ext
|
|
||||||
print("[extract] filename: " + link)
|
|
||||||
b64end = rest.find(")")
|
|
||||||
if b64end != -1:
|
|
||||||
b64encoded = rest[:b64end]
|
|
||||||
print("[extract] %d text bytes" % len(b64encoded))
|
|
||||||
# write if not cached
|
|
||||||
if b64encoded not in cache:
|
|
||||||
try:
|
|
||||||
content = base64.b64decode(b64encoded + "==")
|
|
||||||
open(public + link, "wb").write(content)
|
|
||||||
print("[extract] " + str(len(content)) + " image bytes")
|
|
||||||
cache[b64encoded] = name
|
|
||||||
except Exception:
|
|
||||||
raise Exception
|
|
||||||
# raise Exception('[extract] error decoding image %r' %b64encoded)
|
|
||||||
else:
|
|
||||||
print("[extract] 0 image bytes, cached for " + cache[b64encoded])
|
|
||||||
name = cache[b64encoded]
|
|
||||||
|
|
||||||
# update link with CDN
|
|
||||||
link = cdn + "/upload/image-" + name + "." + ext
|
|
||||||
|
|
||||||
# patch newparts
|
|
||||||
newparts[i + 1] = link + rest[b64end:]
|
|
||||||
else:
|
|
||||||
raise Exception("cannot find the end of base64 encoded string")
|
|
||||||
else:
|
|
||||||
print("[extract] dataimage skipping part " + str(i))
|
|
||||||
continue
|
|
||||||
return "".join(newparts)
|
|
||||||
|
|
||||||
|
|
||||||
di = "data:image"
|
di = "data:image"
|
||||||
|
|
||||||
|
|
||||||
def extract_md_images(body, prefix):
|
|
||||||
newbody = ""
|
|
||||||
body = (
|
|
||||||
body.replace("\n! [](" + di, "\n 
|
|
||||||
.replace("\n[](" + di, "\n
|
|
||||||
.replace(" [](" + di, " 
|
|
||||||
)
|
|
||||||
parts = body.split(di)
|
|
||||||
if len(parts) > 1:
|
|
||||||
newbody = extract_dataimages(parts, prefix)
|
|
||||||
else:
|
|
||||||
newbody = body
|
|
||||||
return newbody
|
|
||||||
|
|
||||||
|
|
||||||
def cleanup_md(body):
|
|
||||||
newbody = (
|
|
||||||
body.replace("<", "")
|
|
||||||
.replace(">", "")
|
|
||||||
.replace("{", "(")
|
|
||||||
.replace("}", ")")
|
|
||||||
.replace("…", "...")
|
|
||||||
.replace(" __ ", " ")
|
|
||||||
.replace("_ _", " ")
|
|
||||||
.replace("****", "")
|
|
||||||
.replace("\u00a0", " ")
|
|
||||||
.replace("\u02c6", "^")
|
|
||||||
.replace("\u00a0", " ")
|
|
||||||
.replace("\ufeff", "")
|
|
||||||
.replace("\u200b", "")
|
|
||||||
.replace("\u200c", "")
|
|
||||||
) # .replace('\u2212', '-')
|
|
||||||
return newbody
|
|
||||||
|
|
||||||
|
|
||||||
def extract_md(body, shout_dict = None):
|
|
||||||
newbody = body
|
|
||||||
if newbody:
|
|
||||||
newbody = cleanup_md(newbody)
|
|
||||||
if not newbody:
|
|
||||||
raise Exception("cleanup error")
|
|
||||||
|
|
||||||
if shout_dict:
|
|
||||||
|
|
||||||
uid = shout_dict['id'] or uuid.uuid4()
|
|
||||||
newbody = extract_md_images(newbody, uid)
|
|
||||||
if not newbody:
|
|
||||||
raise Exception("extract_images error")
|
|
||||||
|
|
||||||
newbody, placed = extract_footnotes(body, shout_dict)
|
|
||||||
if not newbody:
|
|
||||||
raise Exception("extract_footnotes error")
|
|
||||||
|
|
||||||
return newbody
|
|
||||||
|
|
||||||
|
|
||||||
def extract_media(entry):
|
def extract_media(entry):
|
||||||
''' normalized media extraction method '''
|
"""normalized media extraction method"""
|
||||||
# media [ { title pic url body } ]}
|
# media [ { title pic url body } ]}
|
||||||
kind = entry.get("type")
|
kind = entry.get("type")
|
||||||
if not kind:
|
if not kind:
|
||||||
|
@ -323,12 +171,7 @@ def extract_media(entry):
|
||||||
url = "https://vimeo.com/" + m["vimeoId"]
|
url = "https://vimeo.com/" + m["vimeoId"]
|
||||||
# body
|
# body
|
||||||
body = m.get("body") or m.get("literatureBody") or ""
|
body = m.get("body") or m.get("literatureBody") or ""
|
||||||
media.append({
|
media.append({"url": url, "pic": pic, "title": title, "body": body})
|
||||||
"url": url,
|
|
||||||
"pic": pic,
|
|
||||||
"title": title,
|
|
||||||
"body": body
|
|
||||||
})
|
|
||||||
return media
|
return media
|
||||||
|
|
||||||
|
|
||||||
|
@ -398,9 +241,7 @@ def cleanup_html(body: str) -> str:
|
||||||
r"<h4>\s*</h4>",
|
r"<h4>\s*</h4>",
|
||||||
r"<div>\s*</div>",
|
r"<div>\s*</div>",
|
||||||
]
|
]
|
||||||
regex_replace = {
|
regex_replace = {r"<br>\s*</p>": "</p>"}
|
||||||
r"<br>\s*</p>": "</p>"
|
|
||||||
}
|
|
||||||
changed = True
|
changed = True
|
||||||
while changed:
|
while changed:
|
||||||
# we need several iterations to clean nested tags this way
|
# we need several iterations to clean nested tags this way
|
||||||
|
@ -414,16 +255,17 @@ def cleanup_html(body: str) -> str:
|
||||||
changed = True
|
changed = True
|
||||||
return new_body
|
return new_body
|
||||||
|
|
||||||
def extract_html(entry, shout_id = None, cleanup=False):
|
|
||||||
body_orig = (entry.get("body") or "").replace('\(', '(').replace('\)', ')')
|
def extract_html(entry, shout_id=None, cleanup=False):
|
||||||
|
body_orig = (entry.get("body") or "").replace(r"\(", "(").replace(r"\)", ")")
|
||||||
if cleanup:
|
if cleanup:
|
||||||
# we do that before bs parsing to catch the invalid html
|
# we do that before bs parsing to catch the invalid html
|
||||||
body_clean = cleanup_html(body_orig)
|
body_clean = cleanup_html(body_orig)
|
||||||
if body_clean != body_orig:
|
if body_clean != body_orig:
|
||||||
print(f"[migration] html cleaned for slug {entry.get('slug', None)}")
|
print(f"[migration] html cleaned for slug {entry.get('slug', None)}")
|
||||||
body_orig = body_clean
|
body_orig = body_clean
|
||||||
if shout_id:
|
# if shout_id:
|
||||||
extract_footnotes(body_orig, shout_id)
|
# extract_footnotes(body_orig, shout_id)
|
||||||
body_html = str(BeautifulSoup(body_orig, features="html.parser"))
|
body_html = str(BeautifulSoup(body_orig, features="html.parser"))
|
||||||
if cleanup:
|
if cleanup:
|
||||||
# we do that after bs parsing because it can add dummy tags
|
# we do that after bs parsing because it can add dummy tags
|
||||||
|
|
|
@ -33,7 +33,7 @@ __version__ = (2020, 1, 16)
|
||||||
# TODO: Support decoded entities with UNIFIABLE.
|
# TODO: Support decoded entities with UNIFIABLE.
|
||||||
|
|
||||||
|
|
||||||
class HTML2Text(html.parser.HTMLParser):
|
class HTML2Text(html.parser.HTMLParser): # noqa: C901
|
||||||
def __init__(
|
def __init__(
|
||||||
self,
|
self,
|
||||||
out: Optional[OutCallback] = None,
|
out: Optional[OutCallback] = None,
|
||||||
|
@ -85,7 +85,7 @@ class HTML2Text(html.parser.HTMLParser):
|
||||||
self.tag_callback = None
|
self.tag_callback = None
|
||||||
self.open_quote = config.OPEN_QUOTE # covered in cli
|
self.open_quote = config.OPEN_QUOTE # covered in cli
|
||||||
self.close_quote = config.CLOSE_QUOTE # covered in cli
|
self.close_quote = config.CLOSE_QUOTE # covered in cli
|
||||||
self.header_id = None
|
self.header_id: str | None = None
|
||||||
self.span_highlight = False
|
self.span_highlight = False
|
||||||
self.span_lead = False
|
self.span_lead = False
|
||||||
|
|
||||||
|
@ -119,9 +119,7 @@ class HTML2Text(html.parser.HTMLParser):
|
||||||
self.lastWasList = False
|
self.lastWasList = False
|
||||||
self.style = 0
|
self.style = 0
|
||||||
self.style_def = {} # type: Dict[str, Dict[str, str]]
|
self.style_def = {} # type: Dict[str, Dict[str, str]]
|
||||||
self.tag_stack = (
|
self.tag_stack = [] # type: List[Tuple[str, Dict[str, Optional[str]], Dict[str, str]]]
|
||||||
[]
|
|
||||||
) # type: List[Tuple[str, Dict[str, Optional[str]], Dict[str, str]]]
|
|
||||||
self.emphasis = 0
|
self.emphasis = 0
|
||||||
self.drop_white_space = 0
|
self.drop_white_space = 0
|
||||||
self.inheader = False
|
self.inheader = False
|
||||||
|
@ -227,7 +225,7 @@ class HTML2Text(html.parser.HTMLParser):
|
||||||
return i
|
return i
|
||||||
return None
|
return None
|
||||||
|
|
||||||
def handle_emphasis(
|
def handle_emphasis( # noqa: C901
|
||||||
self, start: bool, tag_style: Dict[str, str], parent_style: Dict[str, str]
|
self, start: bool, tag_style: Dict[str, str], parent_style: Dict[str, str]
|
||||||
) -> None:
|
) -> None:
|
||||||
"""
|
"""
|
||||||
|
@ -300,7 +298,7 @@ class HTML2Text(html.parser.HTMLParser):
|
||||||
if strikethrough:
|
if strikethrough:
|
||||||
self.quiet -= 1
|
self.quiet -= 1
|
||||||
|
|
||||||
def handle_tag(
|
def handle_tag( # noqa: C901
|
||||||
self, tag: str, attrs: Dict[str, Optional[str]], start: bool
|
self, tag: str, attrs: Dict[str, Optional[str]], start: bool
|
||||||
) -> None:
|
) -> None:
|
||||||
self.current_tag = tag
|
self.current_tag = tag
|
||||||
|
@ -333,9 +331,7 @@ class HTML2Text(html.parser.HTMLParser):
|
||||||
tag_style = element_style(attrs, self.style_def, parent_style)
|
tag_style = element_style(attrs, self.style_def, parent_style)
|
||||||
self.tag_stack.append((tag, attrs, tag_style))
|
self.tag_stack.append((tag, attrs, tag_style))
|
||||||
else:
|
else:
|
||||||
dummy, attrs, tag_style = (
|
dummy, attrs, tag_style = self.tag_stack.pop() if self.tag_stack else (None, {}, {})
|
||||||
self.tag_stack.pop() if self.tag_stack else (None, {}, {})
|
|
||||||
)
|
|
||||||
if self.tag_stack:
|
if self.tag_stack:
|
||||||
parent_style = self.tag_stack[-1][2]
|
parent_style = self.tag_stack[-1][2]
|
||||||
|
|
||||||
|
@ -385,11 +381,7 @@ class HTML2Text(html.parser.HTMLParser):
|
||||||
):
|
):
|
||||||
self.o("`") # NOTE: same as <code>
|
self.o("`") # NOTE: same as <code>
|
||||||
self.span_highlight = True
|
self.span_highlight = True
|
||||||
elif (
|
elif self.current_class == "lead" and not self.inheader and not self.span_highlight:
|
||||||
self.current_class == "lead"
|
|
||||||
and not self.inheader
|
|
||||||
and not self.span_highlight
|
|
||||||
):
|
|
||||||
# self.o("==") # NOTE: CriticMarkup {==
|
# self.o("==") # NOTE: CriticMarkup {==
|
||||||
self.span_lead = True
|
self.span_lead = True
|
||||||
else:
|
else:
|
||||||
|
@ -479,11 +471,7 @@ class HTML2Text(html.parser.HTMLParser):
|
||||||
and not self.span_lead
|
and not self.span_lead
|
||||||
and not self.span_highlight
|
and not self.span_highlight
|
||||||
):
|
):
|
||||||
if (
|
if start and self.preceding_data and self.preceding_data[-1] == self.strong_mark[0]:
|
||||||
start
|
|
||||||
and self.preceding_data
|
|
||||||
and self.preceding_data[-1] == self.strong_mark[0]
|
|
||||||
):
|
|
||||||
strong = " " + self.strong_mark
|
strong = " " + self.strong_mark
|
||||||
self.preceding_data += " "
|
self.preceding_data += " "
|
||||||
else:
|
else:
|
||||||
|
@ -548,13 +536,8 @@ class HTML2Text(html.parser.HTMLParser):
|
||||||
"href" in attrs
|
"href" in attrs
|
||||||
and not attrs["href"].startswith("#_ftn")
|
and not attrs["href"].startswith("#_ftn")
|
||||||
and attrs["href"] is not None
|
and attrs["href"] is not None
|
||||||
and not (
|
and not (self.skip_internal_links and attrs["href"].startswith("#"))
|
||||||
self.skip_internal_links and attrs["href"].startswith("#")
|
and not (self.ignore_mailto_links and attrs["href"].startswith("mailto:"))
|
||||||
)
|
|
||||||
and not (
|
|
||||||
self.ignore_mailto_links
|
|
||||||
and attrs["href"].startswith("mailto:")
|
|
||||||
)
|
|
||||||
):
|
):
|
||||||
self.astack.append(attrs)
|
self.astack.append(attrs)
|
||||||
self.maybe_automatic_link = attrs["href"]
|
self.maybe_automatic_link = attrs["href"]
|
||||||
|
@ -591,7 +574,7 @@ class HTML2Text(html.parser.HTMLParser):
|
||||||
|
|
||||||
if tag == "img" and start and not self.ignore_images:
|
if tag == "img" and start and not self.ignore_images:
|
||||||
# skip cloudinary images
|
# skip cloudinary images
|
||||||
if "src" in attrs and "cloudinary" not in attrs["src"]:
|
if "src" in attrs and ("cloudinary" not in attrs["src"]):
|
||||||
assert attrs["src"] is not None
|
assert attrs["src"] is not None
|
||||||
if not self.images_to_alt:
|
if not self.images_to_alt:
|
||||||
attrs["href"] = attrs["src"]
|
attrs["href"] = attrs["src"]
|
||||||
|
@ -638,9 +621,7 @@ class HTML2Text(html.parser.HTMLParser):
|
||||||
self.o("![" + escape_md(alt) + "]")
|
self.o("![" + escape_md(alt) + "]")
|
||||||
if self.inline_links:
|
if self.inline_links:
|
||||||
href = attrs.get("href") or ""
|
href = attrs.get("href") or ""
|
||||||
self.o(
|
self.o("(" + escape_md(urlparse.urljoin(self.baseurl, href)) + ")")
|
||||||
"(" + escape_md(urlparse.urljoin(self.baseurl, href)) + ")"
|
|
||||||
)
|
|
||||||
else:
|
else:
|
||||||
i = self.previousIndex(attrs)
|
i = self.previousIndex(attrs)
|
||||||
if i is not None:
|
if i is not None:
|
||||||
|
@ -696,9 +677,7 @@ class HTML2Text(html.parser.HTMLParser):
|
||||||
# WARNING: does not line up <ol><li>s > 9 correctly.
|
# WARNING: does not line up <ol><li>s > 9 correctly.
|
||||||
parent_list = None
|
parent_list = None
|
||||||
for list in self.list:
|
for list in self.list:
|
||||||
self.o(
|
self.o(" " if parent_list == "ol" and list.name == "ul" else " ")
|
||||||
" " if parent_list == "ol" and list.name == "ul" else " "
|
|
||||||
)
|
|
||||||
parent_list = list.name
|
parent_list = list.name
|
||||||
|
|
||||||
if li.name == "ul":
|
if li.name == "ul":
|
||||||
|
@ -787,7 +766,7 @@ class HTML2Text(html.parser.HTMLParser):
|
||||||
self.pbr()
|
self.pbr()
|
||||||
self.br_toggle = " "
|
self.br_toggle = " "
|
||||||
|
|
||||||
def o(
|
def o( # noqa: C901
|
||||||
self, data: str, puredata: bool = False, force: Union[bool, str] = False
|
self, data: str, puredata: bool = False, force: Union[bool, str] = False
|
||||||
) -> None:
|
) -> None:
|
||||||
"""
|
"""
|
||||||
|
@ -864,9 +843,7 @@ class HTML2Text(html.parser.HTMLParser):
|
||||||
self.out(" ")
|
self.out(" ")
|
||||||
self.space = False
|
self.space = False
|
||||||
|
|
||||||
if self.a and (
|
if self.a and ((self.p_p == 2 and self.links_each_paragraph) or force == "end"):
|
||||||
(self.p_p == 2 and self.links_each_paragraph) or force == "end"
|
|
||||||
):
|
|
||||||
if force == "end":
|
if force == "end":
|
||||||
self.out("\n")
|
self.out("\n")
|
||||||
|
|
||||||
|
@ -925,11 +902,7 @@ class HTML2Text(html.parser.HTMLParser):
|
||||||
|
|
||||||
if self.maybe_automatic_link is not None:
|
if self.maybe_automatic_link is not None:
|
||||||
href = self.maybe_automatic_link
|
href = self.maybe_automatic_link
|
||||||
if (
|
if href == data and self.absolute_url_matcher.match(href) and self.use_automatic_links:
|
||||||
href == data
|
|
||||||
and self.absolute_url_matcher.match(href)
|
|
||||||
and self.use_automatic_links
|
|
||||||
):
|
|
||||||
self.o("<" + data + ">")
|
self.o("<" + data + ">")
|
||||||
self.empty_link = False
|
self.empty_link = False
|
||||||
return
|
return
|
||||||
|
@ -980,7 +953,7 @@ class HTML2Text(html.parser.HTMLParser):
|
||||||
|
|
||||||
return nest_count
|
return nest_count
|
||||||
|
|
||||||
def optwrap(self, text: str) -> str:
|
def optwrap(self, text: str) -> str: # noqa: C901
|
||||||
"""
|
"""
|
||||||
Wrap all paragraphs in the provided text.
|
Wrap all paragraphs in the provided text.
|
||||||
|
|
||||||
|
@ -1000,9 +973,7 @@ class HTML2Text(html.parser.HTMLParser):
|
||||||
self.inline_links = False
|
self.inline_links = False
|
||||||
for para in text.split("\n"):
|
for para in text.split("\n"):
|
||||||
if len(para) > 0:
|
if len(para) > 0:
|
||||||
if not skipwrap(
|
if not skipwrap(para, self.wrap_links, self.wrap_list_items, self.wrap_tables):
|
||||||
para, self.wrap_links, self.wrap_list_items, self.wrap_tables
|
|
||||||
):
|
|
||||||
indent = ""
|
indent = ""
|
||||||
if para.startswith(" " + self.ul_item_mark):
|
if para.startswith(" " + self.ul_item_mark):
|
||||||
# list item continuation: add a double indent to the
|
# list item continuation: add a double indent to the
|
||||||
|
@ -1043,12 +1014,10 @@ class HTML2Text(html.parser.HTMLParser):
|
||||||
return result
|
return result
|
||||||
|
|
||||||
|
|
||||||
def html2text(
|
def html2text(html: str, baseurl: str = "", bodywidth: int = config.BODY_WIDTH) -> str:
|
||||||
html: str, baseurl: str = "", bodywidth: Optional[int] = config.BODY_WIDTH
|
|
||||||
) -> str:
|
|
||||||
h = html.strip() or ""
|
h = html.strip() or ""
|
||||||
if h:
|
if h:
|
||||||
h = HTML2Text(baseurl=baseurl, bodywidth=bodywidth)
|
h2t = HTML2Text(baseurl=baseurl, bodywidth=bodywidth)
|
||||||
h = h.handle(html.strip())
|
h = h2t.handle(html.strip())
|
||||||
# print('[html2text] %d bytes' % len(html))
|
# print('[html2text] %d bytes' % len(html))
|
||||||
return h
|
return h
|
||||||
|
|
|
@ -117,10 +117,7 @@ def main() -> None:
|
||||||
dest="images_with_size",
|
dest="images_with_size",
|
||||||
action="store_true",
|
action="store_true",
|
||||||
default=config.IMAGES_WITH_SIZE,
|
default=config.IMAGES_WITH_SIZE,
|
||||||
help=(
|
help=("Write image tags with height and width attrs as raw html to retain " "dimensions"),
|
||||||
"Write image tags with height and width attrs as raw html to retain "
|
|
||||||
"dimensions"
|
|
||||||
),
|
|
||||||
)
|
)
|
||||||
p.add_argument(
|
p.add_argument(
|
||||||
"-g",
|
"-g",
|
||||||
|
@ -260,9 +257,7 @@ def main() -> None:
|
||||||
default=config.CLOSE_QUOTE,
|
default=config.CLOSE_QUOTE,
|
||||||
help="The character used to close quotes",
|
help="The character used to close quotes",
|
||||||
)
|
)
|
||||||
p.add_argument(
|
p.add_argument("--version", action="version", version=".".join(map(str, __version__)))
|
||||||
"--version", action="version", version=".".join(map(str, __version__))
|
|
||||||
)
|
|
||||||
p.add_argument("filename", nargs="?")
|
p.add_argument("filename", nargs="?")
|
||||||
p.add_argument("encoding", nargs="?", default="utf-8")
|
p.add_argument("encoding", nargs="?", default="utf-8")
|
||||||
args = p.parse_args()
|
args = p.parse_args()
|
||||||
|
|
|
@ -4,9 +4,7 @@ from typing import Dict, List, Optional
|
||||||
from . import config
|
from . import config
|
||||||
|
|
||||||
unifiable_n = {
|
unifiable_n = {
|
||||||
html.entities.name2codepoint[k]: v
|
html.entities.name2codepoint[k]: v for k, v in config.UNIFIABLE.items() if k != "nbsp"
|
||||||
for k, v in config.UNIFIABLE.items()
|
|
||||||
if k != "nbsp"
|
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
|
@ -68,12 +66,14 @@ def element_style(
|
||||||
:rtype: dict
|
:rtype: dict
|
||||||
"""
|
"""
|
||||||
style = parent_style.copy()
|
style = parent_style.copy()
|
||||||
if attrs.get("class"):
|
attrs_class = attrs.get("class")
|
||||||
for css_class in attrs["class"].split():
|
if attrs_class:
|
||||||
|
for css_class in attrs_class.split():
|
||||||
css_style = style_def.get("." + css_class, {})
|
css_style = style_def.get("." + css_class, {})
|
||||||
style.update(css_style)
|
style.update(css_style)
|
||||||
if attrs.get("style"):
|
attrs_style = attrs.get("style")
|
||||||
immediate_style = dumb_property_dict(attrs["style"])
|
if attrs_style:
|
||||||
|
immediate_style = dumb_property_dict(attrs_style)
|
||||||
style.update(immediate_style)
|
style.update(immediate_style)
|
||||||
|
|
||||||
return style
|
return style
|
||||||
|
@ -147,18 +147,17 @@ def list_numbering_start(attrs: Dict[str, Optional[str]]) -> int:
|
||||||
|
|
||||||
:rtype: int or None
|
:rtype: int or None
|
||||||
"""
|
"""
|
||||||
if attrs.get("start"):
|
attrs_start = attrs.get("start")
|
||||||
|
if attrs_start:
|
||||||
try:
|
try:
|
||||||
return int(attrs["start"]) - 1
|
return int(attrs_start) - 1
|
||||||
except ValueError:
|
except ValueError:
|
||||||
pass
|
pass
|
||||||
|
|
||||||
return 0
|
return 0
|
||||||
|
|
||||||
|
|
||||||
def skipwrap(
|
def skipwrap(para: str, wrap_links: bool, wrap_list_items: bool, wrap_tables: bool) -> bool:
|
||||||
para: str, wrap_links: bool, wrap_list_items: bool, wrap_tables: bool
|
|
||||||
) -> bool:
|
|
||||||
# If it appears to contain a link
|
# If it appears to contain a link
|
||||||
# don't wrap
|
# don't wrap
|
||||||
if not wrap_links and config.RE_LINK.search(para):
|
if not wrap_links and config.RE_LINK.search(para):
|
||||||
|
@ -236,9 +235,7 @@ def reformat_table(lines: List[str], right_margin: int) -> List[str]:
|
||||||
max_width += [len(x) + right_margin for x in cols[-(num_cols - max_cols) :]]
|
max_width += [len(x) + right_margin for x in cols[-(num_cols - max_cols) :]]
|
||||||
max_cols = num_cols
|
max_cols = num_cols
|
||||||
|
|
||||||
max_width = [
|
max_width = [max(len(x) + right_margin, old_len) for x, old_len in zip(cols, max_width)]
|
||||||
max(len(x) + right_margin, old_len) for x, old_len in zip(cols, max_width)
|
|
||||||
]
|
|
||||||
|
|
||||||
# reformat
|
# reformat
|
||||||
new_lines = []
|
new_lines = []
|
||||||
|
@ -247,15 +244,13 @@ def reformat_table(lines: List[str], right_margin: int) -> List[str]:
|
||||||
if set(line.strip()) == set("-|"):
|
if set(line.strip()) == set("-|"):
|
||||||
filler = "-"
|
filler = "-"
|
||||||
new_cols = [
|
new_cols = [
|
||||||
x.rstrip() + (filler * (M - len(x.rstrip())))
|
x.rstrip() + (filler * (M - len(x.rstrip()))) for x, M in zip(cols, max_width)
|
||||||
for x, M in zip(cols, max_width)
|
|
||||||
]
|
]
|
||||||
new_lines.append("|-" + "|".join(new_cols) + "|")
|
new_lines.append("|-" + "|".join(new_cols) + "|")
|
||||||
else:
|
else:
|
||||||
filler = " "
|
filler = " "
|
||||||
new_cols = [
|
new_cols = [
|
||||||
x.rstrip() + (filler * (M - len(x.rstrip())))
|
x.rstrip() + (filler * (M - len(x.rstrip()))) for x, M in zip(cols, max_width)
|
||||||
for x, M in zip(cols, max_width)
|
|
||||||
]
|
]
|
||||||
new_lines.append("| " + "|".join(new_cols) + "|")
|
new_lines.append("| " + "|".join(new_cols) + "|")
|
||||||
return new_lines
|
return new_lines
|
||||||
|
|
|
@ -1 +0,0 @@
|
||||||
__all__ = (["users", "topics", "content_items", "comments"],)
|
|
|
@ -5,61 +5,48 @@ from dateutil.parser import parse as date_parse
|
||||||
from base.orm import local_session
|
from base.orm import local_session
|
||||||
from migration.html2text import html2text
|
from migration.html2text import html2text
|
||||||
from orm.reaction import Reaction, ReactionKind
|
from orm.reaction import Reaction, ReactionKind
|
||||||
from orm.shout import ShoutReactionsFollower
|
from orm.shout import Shout, ShoutReactionsFollower
|
||||||
from orm.topic import TopicFollower
|
from orm.topic import TopicFollower
|
||||||
from orm.user import User
|
from orm.user import User
|
||||||
from orm.shout import Shout
|
|
||||||
|
|
||||||
ts = datetime.now(tz=timezone.utc)
|
ts = datetime.now(tz=timezone.utc)
|
||||||
|
|
||||||
|
|
||||||
def auto_followers(session, topics, reaction_dict):
|
def auto_followers(session, topics, reaction_dict):
|
||||||
# creating shout's reactions following for reaction author
|
# creating shout's reactions following for reaction author
|
||||||
following1 = session.query(
|
following1 = (
|
||||||
ShoutReactionsFollower
|
session.query(ShoutReactionsFollower)
|
||||||
).where(
|
.where(ShoutReactionsFollower.follower == reaction_dict["createdBy"])
|
||||||
ShoutReactionsFollower.follower == reaction_dict["createdBy"]
|
.filter(ShoutReactionsFollower.shout == reaction_dict["shout"])
|
||||||
).filter(
|
.first()
|
||||||
ShoutReactionsFollower.shout == reaction_dict["shout"]
|
)
|
||||||
).first()
|
|
||||||
if not following1:
|
if not following1:
|
||||||
following1 = ShoutReactionsFollower.create(
|
following1 = ShoutReactionsFollower.create(
|
||||||
follower=reaction_dict["createdBy"],
|
follower=reaction_dict["createdBy"], shout=reaction_dict["shout"], auto=True
|
||||||
shout=reaction_dict["shout"],
|
|
||||||
auto=True
|
|
||||||
)
|
)
|
||||||
session.add(following1)
|
session.add(following1)
|
||||||
# creating topics followings for reaction author
|
# creating topics followings for reaction author
|
||||||
for t in topics:
|
for t in topics:
|
||||||
tf = session.query(
|
tf = (
|
||||||
TopicFollower
|
session.query(TopicFollower)
|
||||||
).where(
|
.where(TopicFollower.follower == reaction_dict["createdBy"])
|
||||||
TopicFollower.follower == reaction_dict["createdBy"]
|
.filter(TopicFollower.topic == t["id"])
|
||||||
).filter(
|
.first()
|
||||||
TopicFollower.topic == t['id']
|
)
|
||||||
).first()
|
|
||||||
if not tf:
|
if not tf:
|
||||||
topic_following = TopicFollower.create(
|
topic_following = TopicFollower.create(
|
||||||
follower=reaction_dict["createdBy"],
|
follower=reaction_dict["createdBy"], topic=t["id"], auto=True
|
||||||
topic=t['id'],
|
|
||||||
auto=True
|
|
||||||
)
|
)
|
||||||
session.add(topic_following)
|
session.add(topic_following)
|
||||||
|
|
||||||
|
|
||||||
def migrate_ratings(session, entry, reaction_dict):
|
def migrate_ratings(session, entry, reaction_dict):
|
||||||
for comment_rating_old in entry.get("ratings", []):
|
for comment_rating_old in entry.get("ratings", []):
|
||||||
rater = (
|
rater = session.query(User).filter(User.oid == comment_rating_old["createdBy"]).first()
|
||||||
session.query(User)
|
|
||||||
.filter(User.oid == comment_rating_old["createdBy"])
|
|
||||||
.first()
|
|
||||||
)
|
|
||||||
re_reaction_dict = {
|
re_reaction_dict = {
|
||||||
"shout": reaction_dict["shout"],
|
"shout": reaction_dict["shout"],
|
||||||
"replyTo": reaction_dict["id"],
|
"replyTo": reaction_dict["id"],
|
||||||
"kind": ReactionKind.LIKE
|
"kind": ReactionKind.LIKE if comment_rating_old["value"] > 0 else ReactionKind.DISLIKE,
|
||||||
if comment_rating_old["value"] > 0
|
|
||||||
else ReactionKind.DISLIKE,
|
|
||||||
"createdBy": rater.id if rater else 1,
|
"createdBy": rater.id if rater else 1,
|
||||||
}
|
}
|
||||||
cts = comment_rating_old.get("createdAt")
|
cts = comment_rating_old.get("createdAt")
|
||||||
|
@ -68,18 +55,15 @@ def migrate_ratings(session, entry, reaction_dict):
|
||||||
try:
|
try:
|
||||||
# creating reaction from old rating
|
# creating reaction from old rating
|
||||||
rr = Reaction.create(**re_reaction_dict)
|
rr = Reaction.create(**re_reaction_dict)
|
||||||
following2 = session.query(
|
following2 = (
|
||||||
ShoutReactionsFollower
|
session.query(ShoutReactionsFollower)
|
||||||
).where(
|
.where(ShoutReactionsFollower.follower == re_reaction_dict["createdBy"])
|
||||||
ShoutReactionsFollower.follower == re_reaction_dict['createdBy']
|
.filter(ShoutReactionsFollower.shout == rr.shout)
|
||||||
).filter(
|
.first()
|
||||||
ShoutReactionsFollower.shout == rr.shout
|
)
|
||||||
).first()
|
|
||||||
if not following2:
|
if not following2:
|
||||||
following2 = ShoutReactionsFollower.create(
|
following2 = ShoutReactionsFollower.create(
|
||||||
follower=re_reaction_dict['createdBy'],
|
follower=re_reaction_dict["createdBy"], shout=rr.shout, auto=True
|
||||||
shout=rr.shout,
|
|
||||||
auto=True
|
|
||||||
)
|
)
|
||||||
session.add(following2)
|
session.add(following2)
|
||||||
session.add(rr)
|
session.add(rr)
|
||||||
|
@ -150,9 +134,7 @@ async def migrate(entry, storage):
|
||||||
else:
|
else:
|
||||||
stage = "author and old id found"
|
stage = "author and old id found"
|
||||||
try:
|
try:
|
||||||
shout = session.query(
|
shout = session.query(Shout).where(Shout.slug == old_shout["slug"]).one()
|
||||||
Shout
|
|
||||||
).where(Shout.slug == old_shout["slug"]).one()
|
|
||||||
if shout:
|
if shout:
|
||||||
reaction_dict["shout"] = shout.id
|
reaction_dict["shout"] = shout.id
|
||||||
reaction_dict["createdBy"] = author.id if author else 1
|
reaction_dict["createdBy"] = author.id if author else 1
|
||||||
|
@ -178,9 +160,9 @@ async def migrate(entry, storage):
|
||||||
|
|
||||||
|
|
||||||
def migrate_2stage(old_comment, idmap):
|
def migrate_2stage(old_comment, idmap):
|
||||||
if old_comment.get('body'):
|
if old_comment.get("body"):
|
||||||
new_id = idmap.get(old_comment.get('oid'))
|
new_id = idmap.get(old_comment.get("oid"))
|
||||||
new_id = idmap.get(old_comment.get('_id'))
|
new_id = idmap.get(old_comment.get("_id"))
|
||||||
if new_id:
|
if new_id:
|
||||||
new_replyto_id = None
|
new_replyto_id = None
|
||||||
old_replyto_id = old_comment.get("replyTo")
|
old_replyto_id = old_comment.get("replyTo")
|
||||||
|
@ -190,17 +172,20 @@ def migrate_2stage(old_comment, idmap):
|
||||||
comment = session.query(Reaction).where(Reaction.id == new_id).first()
|
comment = session.query(Reaction).where(Reaction.id == new_id).first()
|
||||||
try:
|
try:
|
||||||
if new_replyto_id:
|
if new_replyto_id:
|
||||||
new_reply = session.query(Reaction).where(Reaction.id == new_replyto_id).first()
|
new_reply = (
|
||||||
|
session.query(Reaction).where(Reaction.id == new_replyto_id).first()
|
||||||
|
)
|
||||||
if not new_reply:
|
if not new_reply:
|
||||||
print(new_replyto_id)
|
print(new_replyto_id)
|
||||||
raise Exception("cannot find reply by id!")
|
raise Exception("cannot find reply by id!")
|
||||||
comment.replyTo = new_reply.id
|
comment.replyTo = new_reply.id
|
||||||
session.add(comment)
|
session.add(comment)
|
||||||
srf = session.query(ShoutReactionsFollower).where(
|
srf = (
|
||||||
ShoutReactionsFollower.shout == comment.shout
|
session.query(ShoutReactionsFollower)
|
||||||
).filter(
|
.where(ShoutReactionsFollower.shout == comment.shout)
|
||||||
ShoutReactionsFollower.follower == comment.createdBy
|
.filter(ShoutReactionsFollower.follower == comment.createdBy)
|
||||||
).first()
|
.first()
|
||||||
|
)
|
||||||
if not srf:
|
if not srf:
|
||||||
srf = ShoutReactionsFollower.create(
|
srf = ShoutReactionsFollower.create(
|
||||||
shout=comment.shout, follower=comment.createdBy, auto=True
|
shout=comment.shout, follower=comment.createdBy, auto=True
|
||||||
|
|
|
@ -1,16 +1,18 @@
|
||||||
from datetime import datetime, timezone
|
|
||||||
import json
|
import json
|
||||||
|
import re
|
||||||
|
from datetime import datetime, timezone
|
||||||
|
|
||||||
from dateutil.parser import parse as date_parse
|
from dateutil.parser import parse as date_parse
|
||||||
from sqlalchemy.exc import IntegrityError
|
from sqlalchemy.exc import IntegrityError
|
||||||
from transliterate import translit
|
from transliterate import translit
|
||||||
|
|
||||||
from base.orm import local_session
|
from base.orm import local_session
|
||||||
from migration.extract import extract_html, extract_media
|
from migration.extract import extract_html, extract_media
|
||||||
from orm.reaction import Reaction, ReactionKind
|
from orm.reaction import Reaction, ReactionKind
|
||||||
from orm.shout import Shout, ShoutTopic, ShoutReactionsFollower
|
from orm.shout import Shout, ShoutReactionsFollower, ShoutTopic
|
||||||
|
from orm.topic import Topic, TopicFollower
|
||||||
from orm.user import User
|
from orm.user import User
|
||||||
from orm.topic import TopicFollower, Topic
|
|
||||||
from services.stat.viewed import ViewedStorage
|
from services.stat.viewed import ViewedStorage
|
||||||
import re
|
|
||||||
|
|
||||||
OLD_DATE = "2016-03-05 22:22:00.350000"
|
OLD_DATE = "2016-03-05 22:22:00.350000"
|
||||||
ts = datetime.now(tz=timezone.utc)
|
ts = datetime.now(tz=timezone.utc)
|
||||||
|
@ -33,7 +35,7 @@ def get_shout_slug(entry):
|
||||||
slug = friend.get("slug", "")
|
slug = friend.get("slug", "")
|
||||||
if slug:
|
if slug:
|
||||||
break
|
break
|
||||||
slug = re.sub('[^0-9a-zA-Z]+', '-', slug)
|
slug = re.sub("[^0-9a-zA-Z]+", "-", slug)
|
||||||
return slug
|
return slug
|
||||||
|
|
||||||
|
|
||||||
|
@ -41,27 +43,27 @@ def create_author_from_app(app):
|
||||||
user = None
|
user = None
|
||||||
userdata = None
|
userdata = None
|
||||||
# check if email is used
|
# check if email is used
|
||||||
if app['email']:
|
if app["email"]:
|
||||||
with local_session() as session:
|
with local_session() as session:
|
||||||
user = session.query(User).where(User.email == app['email']).first()
|
user = session.query(User).where(User.email == app["email"]).first()
|
||||||
if not user:
|
if not user:
|
||||||
# print('[migration] app %r' % app)
|
# print('[migration] app %r' % app)
|
||||||
name = app.get('name')
|
name = app.get("name")
|
||||||
if name:
|
if name:
|
||||||
slug = translit(name, "ru", reversed=True).lower()
|
slug = translit(name, "ru", reversed=True).lower()
|
||||||
slug = re.sub('[^0-9a-zA-Z]+', '-', slug)
|
slug = re.sub("[^0-9a-zA-Z]+", "-", slug)
|
||||||
print('[migration] created slug %s' % slug)
|
print("[migration] created slug %s" % slug)
|
||||||
# check if slug is used
|
# check if slug is used
|
||||||
if slug:
|
if slug:
|
||||||
user = session.query(User).where(User.slug == slug).first()
|
user = session.query(User).where(User.slug == slug).first()
|
||||||
|
|
||||||
# get slug from email
|
# get slug from email
|
||||||
if user:
|
if user:
|
||||||
slug = app['email'].split('@')[0]
|
slug = app["email"].split("@")[0]
|
||||||
user = session.query(User).where(User.slug == slug).first()
|
user = session.query(User).where(User.slug == slug).first()
|
||||||
# one more try
|
# one more try
|
||||||
if user:
|
if user:
|
||||||
slug += '-author'
|
slug += "-author"
|
||||||
user = session.query(User).where(User.slug == slug).first()
|
user = session.query(User).where(User.slug == slug).first()
|
||||||
|
|
||||||
# create user with application data
|
# create user with application data
|
||||||
|
@ -79,7 +81,7 @@ def create_author_from_app(app):
|
||||||
user = User.create(**userdata)
|
user = User.create(**userdata)
|
||||||
session.add(user)
|
session.add(user)
|
||||||
session.commit()
|
session.commit()
|
||||||
userdata['id'] = user.id
|
userdata["id"] = user.id
|
||||||
|
|
||||||
userdata = user.dict()
|
userdata = user.dict()
|
||||||
return userdata
|
return userdata
|
||||||
|
@ -91,11 +93,12 @@ async def create_shout(shout_dict):
|
||||||
s = Shout.create(**shout_dict)
|
s = Shout.create(**shout_dict)
|
||||||
author = s.authors[0]
|
author = s.authors[0]
|
||||||
with local_session() as session:
|
with local_session() as session:
|
||||||
srf = session.query(ShoutReactionsFollower).where(
|
srf = (
|
||||||
ShoutReactionsFollower.shout == s.id
|
session.query(ShoutReactionsFollower)
|
||||||
).filter(
|
.where(ShoutReactionsFollower.shout == s.id)
|
||||||
ShoutReactionsFollower.follower == author.id
|
.filter(ShoutReactionsFollower.follower == author.id)
|
||||||
).first()
|
.first()
|
||||||
|
)
|
||||||
if not srf:
|
if not srf:
|
||||||
srf = ShoutReactionsFollower.create(shout=s.id, follower=author.id, auto=True)
|
srf = ShoutReactionsFollower.create(shout=s.id, follower=author.id, auto=True)
|
||||||
session.add(srf)
|
session.add(srf)
|
||||||
|
@ -116,14 +119,14 @@ async def get_user(entry, storage):
|
||||||
elif user_oid:
|
elif user_oid:
|
||||||
userdata = storage["users"]["by_oid"].get(user_oid)
|
userdata = storage["users"]["by_oid"].get(user_oid)
|
||||||
if not userdata:
|
if not userdata:
|
||||||
print('no userdata by oid, anonymous')
|
print("no userdata by oid, anonymous")
|
||||||
userdata = anondict
|
userdata = anondict
|
||||||
print(app)
|
print(app)
|
||||||
# cleanup slug
|
# cleanup slug
|
||||||
if userdata:
|
if userdata:
|
||||||
slug = userdata.get("slug", "")
|
slug = userdata.get("slug", "")
|
||||||
if slug:
|
if slug:
|
||||||
slug = re.sub('[^0-9a-zA-Z]+', '-', slug)
|
slug = re.sub("[^0-9a-zA-Z]+", "-", slug)
|
||||||
userdata["slug"] = slug
|
userdata["slug"] = slug
|
||||||
else:
|
else:
|
||||||
userdata = anondict
|
userdata = anondict
|
||||||
|
@ -137,11 +140,14 @@ async def migrate(entry, storage):
|
||||||
r = {
|
r = {
|
||||||
"layout": type2layout[entry["type"]],
|
"layout": type2layout[entry["type"]],
|
||||||
"title": entry["title"],
|
"title": entry["title"],
|
||||||
"authors": [author, ],
|
"authors": [
|
||||||
|
author,
|
||||||
|
],
|
||||||
"slug": get_shout_slug(entry),
|
"slug": get_shout_slug(entry),
|
||||||
"cover": (
|
"cover": (
|
||||||
"https://images.discours.io/unsafe/" +
|
"https://images.discours.io/unsafe/" + entry["thumborId"]
|
||||||
entry["thumborId"] if entry.get("thumborId") else entry.get("image", {}).get("url")
|
if entry.get("thumborId")
|
||||||
|
else entry.get("image", {}).get("url")
|
||||||
),
|
),
|
||||||
"visibility": "public" if entry.get("published") else "community",
|
"visibility": "public" if entry.get("published") else "community",
|
||||||
"publishedAt": date_parse(entry.get("publishedAt")) if entry.get("published") else None,
|
"publishedAt": date_parse(entry.get("publishedAt")) if entry.get("published") else None,
|
||||||
|
@ -150,11 +156,11 @@ async def migrate(entry, storage):
|
||||||
"updatedAt": date_parse(entry["updatedAt"]) if "updatedAt" in entry else ts,
|
"updatedAt": date_parse(entry["updatedAt"]) if "updatedAt" in entry else ts,
|
||||||
"createdBy": author.id,
|
"createdBy": author.id,
|
||||||
"topics": await add_topics_follower(entry, storage, author),
|
"topics": await add_topics_follower(entry, storage, author),
|
||||||
"body": extract_html(entry, cleanup=True)
|
"body": extract_html(entry, cleanup=True),
|
||||||
}
|
}
|
||||||
|
|
||||||
# main topic patch
|
# main topic patch
|
||||||
r['mainTopic'] = r['topics'][0]
|
r["mainTopic"] = r["topics"][0]
|
||||||
|
|
||||||
# published author auto-confirm
|
# published author auto-confirm
|
||||||
if entry.get("published"):
|
if entry.get("published"):
|
||||||
|
@ -177,14 +183,16 @@ async def migrate(entry, storage):
|
||||||
shout_dict["oid"] = entry.get("_id", "")
|
shout_dict["oid"] = entry.get("_id", "")
|
||||||
shout = await create_shout(shout_dict)
|
shout = await create_shout(shout_dict)
|
||||||
except IntegrityError as e:
|
except IntegrityError as e:
|
||||||
print('[migration] create_shout integrity error', e)
|
print("[migration] create_shout integrity error", e)
|
||||||
shout = await resolve_create_shout(shout_dict)
|
shout = await resolve_create_shout(shout_dict)
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
raise Exception(e)
|
raise Exception(e)
|
||||||
|
|
||||||
# udpate data
|
# udpate data
|
||||||
shout_dict = shout.dict()
|
shout_dict = shout.dict()
|
||||||
shout_dict["authors"] = [author.dict(), ]
|
shout_dict["authors"] = [
|
||||||
|
author.dict(),
|
||||||
|
]
|
||||||
|
|
||||||
# shout topics aftermath
|
# shout topics aftermath
|
||||||
shout_dict["topics"] = await topics_aftermath(r, storage)
|
shout_dict["topics"] = await topics_aftermath(r, storage)
|
||||||
|
@ -193,7 +201,9 @@ async def migrate(entry, storage):
|
||||||
await content_ratings_to_reactions(entry, shout_dict["slug"])
|
await content_ratings_to_reactions(entry, shout_dict["slug"])
|
||||||
|
|
||||||
# shout views
|
# shout views
|
||||||
await ViewedStorage.increment(shout_dict["slug"], amount=entry.get("views", 1), viewer='old-discours')
|
await ViewedStorage.increment(
|
||||||
|
shout_dict["slug"], amount=entry.get("views", 1), viewer="old-discours"
|
||||||
|
)
|
||||||
# del shout_dict['ratings']
|
# del shout_dict['ratings']
|
||||||
|
|
||||||
storage["shouts"]["by_oid"][entry["_id"]] = shout_dict
|
storage["shouts"]["by_oid"][entry["_id"]] = shout_dict
|
||||||
|
@ -205,7 +215,9 @@ async def add_topics_follower(entry, storage, user):
|
||||||
topics = set([])
|
topics = set([])
|
||||||
category = entry.get("category")
|
category = entry.get("category")
|
||||||
topics_by_oid = storage["topics"]["by_oid"]
|
topics_by_oid = storage["topics"]["by_oid"]
|
||||||
oids = [category, ] + entry.get("tags", [])
|
oids = [
|
||||||
|
category,
|
||||||
|
] + entry.get("tags", [])
|
||||||
for toid in oids:
|
for toid in oids:
|
||||||
tslug = topics_by_oid.get(toid, {}).get("slug")
|
tslug = topics_by_oid.get(toid, {}).get("slug")
|
||||||
if tslug:
|
if tslug:
|
||||||
|
@ -217,23 +229,18 @@ async def add_topics_follower(entry, storage, user):
|
||||||
try:
|
try:
|
||||||
tpc = session.query(Topic).where(Topic.slug == tpcslug).first()
|
tpc = session.query(Topic).where(Topic.slug == tpcslug).first()
|
||||||
if tpc:
|
if tpc:
|
||||||
tf = session.query(
|
tf = (
|
||||||
TopicFollower
|
session.query(TopicFollower)
|
||||||
).where(
|
.where(TopicFollower.follower == user.id)
|
||||||
TopicFollower.follower == user.id
|
.filter(TopicFollower.topic == tpc.id)
|
||||||
).filter(
|
.first()
|
||||||
TopicFollower.topic == tpc.id
|
)
|
||||||
).first()
|
|
||||||
if not tf:
|
if not tf:
|
||||||
tf = TopicFollower.create(
|
tf = TopicFollower.create(topic=tpc.id, follower=user.id, auto=True)
|
||||||
topic=tpc.id,
|
|
||||||
follower=user.id,
|
|
||||||
auto=True
|
|
||||||
)
|
|
||||||
session.add(tf)
|
session.add(tf)
|
||||||
session.commit()
|
session.commit()
|
||||||
except IntegrityError:
|
except IntegrityError:
|
||||||
print('[migration.shout] hidden by topic ' + tpc.slug)
|
print("[migration.shout] hidden by topic " + tpc.slug)
|
||||||
# main topic
|
# main topic
|
||||||
maintopic = storage["replacements"].get(topics_by_oid.get(category, {}).get("slug"))
|
maintopic = storage["replacements"].get(topics_by_oid.get(category, {}).get("slug"))
|
||||||
if maintopic in ttt:
|
if maintopic in ttt:
|
||||||
|
@ -254,7 +261,7 @@ async def process_user(userdata, storage, oid):
|
||||||
if not user:
|
if not user:
|
||||||
try:
|
try:
|
||||||
slug = userdata["slug"].lower().strip()
|
slug = userdata["slug"].lower().strip()
|
||||||
slug = re.sub('[^0-9a-zA-Z]+', '-', slug)
|
slug = re.sub("[^0-9a-zA-Z]+", "-", slug)
|
||||||
userdata["slug"] = slug
|
userdata["slug"] = slug
|
||||||
user = User.create(**userdata)
|
user = User.create(**userdata)
|
||||||
session.add(user)
|
session.add(user)
|
||||||
|
@ -282,9 +289,9 @@ async def resolve_create_shout(shout_dict):
|
||||||
s = session.query(Shout).filter(Shout.slug == shout_dict["slug"]).first()
|
s = session.query(Shout).filter(Shout.slug == shout_dict["slug"]).first()
|
||||||
bump = False
|
bump = False
|
||||||
if s:
|
if s:
|
||||||
if s.createdAt != shout_dict['createdAt']:
|
if s.createdAt != shout_dict["createdAt"]:
|
||||||
# create new with different slug
|
# create new with different slug
|
||||||
shout_dict["slug"] += '-' + shout_dict["layout"]
|
shout_dict["slug"] += "-" + shout_dict["layout"]
|
||||||
try:
|
try:
|
||||||
await create_shout(shout_dict)
|
await create_shout(shout_dict)
|
||||||
except IntegrityError as e:
|
except IntegrityError as e:
|
||||||
|
@ -295,10 +302,7 @@ async def resolve_create_shout(shout_dict):
|
||||||
for key in shout_dict:
|
for key in shout_dict:
|
||||||
if key in s.__dict__:
|
if key in s.__dict__:
|
||||||
if s.__dict__[key] != shout_dict[key]:
|
if s.__dict__[key] != shout_dict[key]:
|
||||||
print(
|
print("[migration] shout already exists, but differs in %s" % key)
|
||||||
"[migration] shout already exists, but differs in %s"
|
|
||||||
% key
|
|
||||||
)
|
|
||||||
bump = True
|
bump = True
|
||||||
else:
|
else:
|
||||||
print("[migration] shout already exists, but lacks %s" % key)
|
print("[migration] shout already exists, but lacks %s" % key)
|
||||||
|
@ -344,9 +348,7 @@ async def topics_aftermath(entry, storage):
|
||||||
)
|
)
|
||||||
if not shout_topic_new:
|
if not shout_topic_new:
|
||||||
try:
|
try:
|
||||||
ShoutTopic.create(
|
ShoutTopic.create(**{"shout": shout.id, "topic": new_topic.id})
|
||||||
**{"shout": shout.id, "topic": new_topic.id}
|
|
||||||
)
|
|
||||||
except Exception:
|
except Exception:
|
||||||
print("[migration] shout topic error: " + newslug)
|
print("[migration] shout topic error: " + newslug)
|
||||||
session.commit()
|
session.commit()
|
||||||
|
@ -363,9 +365,7 @@ async def content_ratings_to_reactions(entry, slug):
|
||||||
with local_session() as session:
|
with local_session() as session:
|
||||||
for content_rating in entry.get("ratings", []):
|
for content_rating in entry.get("ratings", []):
|
||||||
rater = (
|
rater = (
|
||||||
session.query(User)
|
session.query(User).filter(User.oid == content_rating["createdBy"]).first()
|
||||||
.filter(User.oid == content_rating["createdBy"])
|
|
||||||
.first()
|
|
||||||
) or User.default_user
|
) or User.default_user
|
||||||
shout = session.query(Shout).where(Shout.slug == slug).first()
|
shout = session.query(Shout).where(Shout.slug == slug).first()
|
||||||
cts = content_rating.get("createdAt")
|
cts = content_rating.get("createdAt")
|
||||||
|
@ -375,7 +375,7 @@ async def content_ratings_to_reactions(entry, slug):
|
||||||
if content_rating["value"] > 0
|
if content_rating["value"] > 0
|
||||||
else ReactionKind.DISLIKE,
|
else ReactionKind.DISLIKE,
|
||||||
"createdBy": rater.id,
|
"createdBy": rater.id,
|
||||||
"shout": shout.id
|
"shout": shout.id,
|
||||||
}
|
}
|
||||||
reaction = (
|
reaction = (
|
||||||
session.query(Reaction)
|
session.query(Reaction)
|
||||||
|
|
|
@ -1,42 +1,35 @@
|
||||||
from base.orm import local_session
|
# from base.orm import local_session
|
||||||
from migration.extract import extract_md
|
|
||||||
from migration.html2text import html2text
|
# from migration.extract import extract_md
|
||||||
from orm.reaction import Reaction, ReactionKind
|
# from migration.html2text import html2text
|
||||||
|
# from orm.reaction import Reaction, ReactionKind
|
||||||
|
|
||||||
|
|
||||||
def migrate(entry, storage):
|
# def migrate(entry, storage):
|
||||||
post_oid = entry['contentItem']
|
# post_oid = entry["contentItem"]
|
||||||
print(post_oid)
|
# print(post_oid)
|
||||||
shout_dict = storage['shouts']['by_oid'].get(post_oid)
|
# shout_dict = storage["shouts"]["by_oid"].get(post_oid)
|
||||||
if shout_dict:
|
# if shout_dict:
|
||||||
print(shout_dict['body'])
|
# print(shout_dict["body"])
|
||||||
remark = {
|
# remark = {
|
||||||
"shout": shout_dict['id'],
|
# "shout": shout_dict["id"],
|
||||||
"body": extract_md(
|
# "body": extract_md(html2text(entry["body"]), shout_dict),
|
||||||
html2text(entry['body']),
|
# "kind": ReactionKind.REMARK,
|
||||||
shout_dict
|
# }
|
||||||
),
|
#
|
||||||
"kind": ReactionKind.REMARK
|
# if entry.get("textBefore"):
|
||||||
}
|
# remark["range"] = (
|
||||||
|
# str(shout_dict["body"].index(entry["textBefore"] or ""))
|
||||||
if entry.get('textBefore'):
|
# + ":"
|
||||||
remark['range'] = str(
|
# + str(
|
||||||
shout_dict['body']
|
# shout_dict["body"].index(entry["textAfter"] or "")
|
||||||
.index(
|
# + len(entry["textAfter"] or "")
|
||||||
entry['textBefore'] or ''
|
# )
|
||||||
)
|
# )
|
||||||
) + ':' + str(
|
#
|
||||||
shout_dict['body']
|
# with local_session() as session:
|
||||||
.index(
|
# rmrk = Reaction.create(**remark)
|
||||||
entry['textAfter'] or ''
|
# session.commit()
|
||||||
) + len(
|
# del rmrk["_sa_instance_state"]
|
||||||
entry['textAfter'] or ''
|
# return rmrk
|
||||||
)
|
# return
|
||||||
)
|
|
||||||
|
|
||||||
with local_session() as session:
|
|
||||||
rmrk = Reaction.create(**remark)
|
|
||||||
session.commit()
|
|
||||||
del rmrk["_sa_instance_state"]
|
|
||||||
return rmrk
|
|
||||||
return
|
|
||||||
|
|
|
@ -1,5 +1,4 @@
|
||||||
from base.orm import local_session
|
from base.orm import local_session
|
||||||
from migration.extract import extract_md
|
|
||||||
from migration.html2text import html2text
|
from migration.html2text import html2text
|
||||||
from orm import Topic
|
from orm import Topic
|
||||||
|
|
||||||
|
@ -10,7 +9,7 @@ def migrate(entry):
|
||||||
"slug": entry["slug"],
|
"slug": entry["slug"],
|
||||||
"oid": entry["_id"],
|
"oid": entry["_id"],
|
||||||
"title": entry["title"].replace(" ", " "),
|
"title": entry["title"].replace(" ", " "),
|
||||||
"body": extract_md(html2text(body_orig))
|
"body": html2text(body_orig),
|
||||||
}
|
}
|
||||||
|
|
||||||
with local_session() as session:
|
with local_session() as session:
|
||||||
|
|
|
@ -8,7 +8,7 @@ from base.orm import local_session
|
||||||
from orm.user import AuthorFollower, User, UserRating
|
from orm.user import AuthorFollower, User, UserRating
|
||||||
|
|
||||||
|
|
||||||
def migrate(entry):
|
def migrate(entry): # noqa: C901
|
||||||
if "subscribedTo" in entry:
|
if "subscribedTo" in entry:
|
||||||
del entry["subscribedTo"]
|
del entry["subscribedTo"]
|
||||||
email = entry["emails"][0]["address"]
|
email = entry["emails"][0]["address"]
|
||||||
|
@ -23,7 +23,7 @@ def migrate(entry):
|
||||||
"muted": False, # amnesty
|
"muted": False, # amnesty
|
||||||
"links": [],
|
"links": [],
|
||||||
"name": "anonymous",
|
"name": "anonymous",
|
||||||
"password": entry["services"]["password"].get("bcrypt")
|
"password": entry["services"]["password"].get("bcrypt"),
|
||||||
}
|
}
|
||||||
|
|
||||||
if "updatedAt" in entry:
|
if "updatedAt" in entry:
|
||||||
|
@ -33,9 +33,13 @@ def migrate(entry):
|
||||||
if entry.get("profile"):
|
if entry.get("profile"):
|
||||||
# slug
|
# slug
|
||||||
slug = entry["profile"].get("path").lower()
|
slug = entry["profile"].get("path").lower()
|
||||||
slug = re.sub('[^0-9a-zA-Z]+', '-', slug).strip()
|
slug = re.sub("[^0-9a-zA-Z]+", "-", slug).strip()
|
||||||
user_dict["slug"] = slug
|
user_dict["slug"] = slug
|
||||||
bio = (entry.get("profile", {"bio": ""}).get("bio") or "").replace('\(', '(').replace('\)', ')')
|
bio = (
|
||||||
|
(entry.get("profile", {"bio": ""}).get("bio") or "")
|
||||||
|
.replace(r"\(", "(")
|
||||||
|
.replace(r"\)", ")")
|
||||||
|
)
|
||||||
bio_text = BeautifulSoup(bio, features="lxml").text
|
bio_text = BeautifulSoup(bio, features="lxml").text
|
||||||
|
|
||||||
if len(bio_text) > 120:
|
if len(bio_text) > 120:
|
||||||
|
@ -46,8 +50,7 @@ def migrate(entry):
|
||||||
# userpic
|
# userpic
|
||||||
try:
|
try:
|
||||||
user_dict["userpic"] = (
|
user_dict["userpic"] = (
|
||||||
"https://images.discours.io/unsafe/"
|
"https://images.discours.io/unsafe/" + entry["profile"]["thumborId"]
|
||||||
+ entry["profile"]["thumborId"]
|
|
||||||
)
|
)
|
||||||
except KeyError:
|
except KeyError:
|
||||||
try:
|
try:
|
||||||
|
@ -62,11 +65,7 @@ def migrate(entry):
|
||||||
name = (name + " " + ln) if ln else name
|
name = (name + " " + ln) if ln else name
|
||||||
if not name:
|
if not name:
|
||||||
name = slug if slug else "anonymous"
|
name = slug if slug else "anonymous"
|
||||||
name = (
|
name = entry["profile"]["path"].lower().strip().replace(" ", "-") if len(name) < 2 else name
|
||||||
entry["profile"]["path"].lower().strip().replace(" ", "-")
|
|
||||||
if len(name) < 2
|
|
||||||
else name
|
|
||||||
)
|
|
||||||
user_dict["name"] = name
|
user_dict["name"] = name
|
||||||
|
|
||||||
# links
|
# links
|
||||||
|
@ -95,9 +94,7 @@ def migrate(entry):
|
||||||
except IntegrityError:
|
except IntegrityError:
|
||||||
print("[migration] cannot create user " + user_dict["slug"])
|
print("[migration] cannot create user " + user_dict["slug"])
|
||||||
with local_session() as session:
|
with local_session() as session:
|
||||||
old_user = (
|
old_user = session.query(User).filter(User.slug == user_dict["slug"]).first()
|
||||||
session.query(User).filter(User.slug == user_dict["slug"]).first()
|
|
||||||
)
|
|
||||||
old_user.oid = oid
|
old_user.oid = oid
|
||||||
old_user.password = user_dict["password"]
|
old_user.password = user_dict["password"]
|
||||||
session.commit()
|
session.commit()
|
||||||
|
@ -114,7 +111,7 @@ def post_migrate():
|
||||||
"slug": "old-discours",
|
"slug": "old-discours",
|
||||||
"username": "old-discours",
|
"username": "old-discours",
|
||||||
"email": "old@discours.io",
|
"email": "old@discours.io",
|
||||||
"name": "Просмотры на старой версии сайта"
|
"name": "Просмотры на старой версии сайта",
|
||||||
}
|
}
|
||||||
|
|
||||||
with local_session() as session:
|
with local_session() as session:
|
||||||
|
@ -147,12 +144,8 @@ def migrate_2stage(entry, id_map):
|
||||||
}
|
}
|
||||||
|
|
||||||
user_rating = UserRating.create(**user_rating_dict)
|
user_rating = UserRating.create(**user_rating_dict)
|
||||||
if user_rating_dict['value'] > 0:
|
if user_rating_dict["value"] > 0:
|
||||||
af = AuthorFollower.create(
|
af = AuthorFollower.create(author=user.id, follower=rater.id, auto=True)
|
||||||
author=user.id,
|
|
||||||
follower=rater.id,
|
|
||||||
auto=True
|
|
||||||
)
|
|
||||||
session.add(af)
|
session.add(af)
|
||||||
session.add(user_rating)
|
session.add(user_rating)
|
||||||
session.commit()
|
session.commit()
|
||||||
|
|
|
@ -1,7 +1,7 @@
|
||||||
from base.orm import Base, engine
|
from base.orm import Base, engine
|
||||||
from orm.community import Community
|
from orm.community import Community
|
||||||
from orm.notification import Notification
|
from orm.notification import Notification
|
||||||
from orm.rbac import Operation, Resource, Permission, Role
|
from orm.rbac import Operation, Permission, Resource, Role
|
||||||
from orm.reaction import Reaction
|
from orm.reaction import Reaction
|
||||||
from orm.shout import Shout
|
from orm.shout import Shout
|
||||||
from orm.topic import Topic, TopicFollower
|
from orm.topic import Topic, TopicFollower
|
||||||
|
@ -32,5 +32,5 @@ __all__ = [
|
||||||
"Notification",
|
"Notification",
|
||||||
"Reaction",
|
"Reaction",
|
||||||
"UserRating",
|
"UserRating",
|
||||||
"init_tables"
|
"init_tables",
|
||||||
]
|
]
|
||||||
|
|
|
@ -8,7 +8,7 @@ from base.orm import Base
|
||||||
class ShoutCollection(Base):
|
class ShoutCollection(Base):
|
||||||
__tablename__ = "shout_collection"
|
__tablename__ = "shout_collection"
|
||||||
|
|
||||||
id = None # type: ignore
|
id = None
|
||||||
shout = Column(ForeignKey("shout.id"), primary_key=True)
|
shout = Column(ForeignKey("shout.id"), primary_key=True)
|
||||||
collection = Column(ForeignKey("collection.id"), primary_key=True)
|
collection = Column(ForeignKey("collection.id"), primary_key=True)
|
||||||
|
|
||||||
|
|
|
@ -1,18 +1,17 @@
|
||||||
from datetime import datetime
|
from datetime import datetime
|
||||||
|
|
||||||
from sqlalchemy import Column, String, ForeignKey, DateTime
|
from sqlalchemy import Column, DateTime, ForeignKey, String
|
||||||
|
|
||||||
from base.orm import Base, local_session
|
from base.orm import Base, local_session
|
||||||
|
|
||||||
|
|
||||||
class CommunityFollower(Base):
|
class CommunityFollower(Base):
|
||||||
__tablename__ = "community_followers"
|
__tablename__ = "community_followers"
|
||||||
|
|
||||||
id = None # type: ignore
|
id = None
|
||||||
follower = Column(ForeignKey("user.id"), primary_key=True)
|
follower: Column = Column(ForeignKey("user.id"), primary_key=True)
|
||||||
community = Column(ForeignKey("community.id"), primary_key=True)
|
community: Column = Column(ForeignKey("community.id"), primary_key=True)
|
||||||
joinedAt = Column(
|
joinedAt = Column(DateTime, nullable=False, default=datetime.now, comment="Created at")
|
||||||
DateTime, nullable=False, default=datetime.now, comment="Created at"
|
|
||||||
)
|
|
||||||
# role = Column(ForeignKey(Role.id), nullable=False, comment="Role for member")
|
# role = Column(ForeignKey(Role.id), nullable=False, comment="Role for member")
|
||||||
|
|
||||||
|
|
||||||
|
@ -23,19 +22,15 @@ class Community(Base):
|
||||||
slug = Column(String, nullable=False, unique=True, comment="Slug")
|
slug = Column(String, nullable=False, unique=True, comment="Slug")
|
||||||
desc = Column(String, nullable=False, default="")
|
desc = Column(String, nullable=False, default="")
|
||||||
pic = Column(String, nullable=False, default="")
|
pic = Column(String, nullable=False, default="")
|
||||||
createdAt = Column(
|
createdAt = Column(DateTime, nullable=False, default=datetime.now, comment="Created at")
|
||||||
DateTime, nullable=False, default=datetime.now, comment="Created at"
|
|
||||||
)
|
|
||||||
|
|
||||||
@staticmethod
|
@staticmethod
|
||||||
def init_table():
|
def init_table():
|
||||||
with local_session() as session:
|
with local_session() as session:
|
||||||
d = (
|
d = session.query(Community).filter(Community.slug == "discours").first()
|
||||||
session.query(Community).filter(Community.slug == "discours").first()
|
|
||||||
)
|
|
||||||
if not d:
|
if not d:
|
||||||
d = Community.create(name="Дискурс", slug="discours")
|
d = Community.create(name="Дискурс", slug="discours")
|
||||||
session.add(d)
|
session.add(d)
|
||||||
session.commit()
|
session.commit()
|
||||||
Community.default_community = d
|
Community.default_community = d
|
||||||
print('[orm] default community id: %s' % d.id)
|
print("[orm] default community id: %s" % d.id)
|
||||||
|
|
|
@ -1,9 +1,10 @@
|
||||||
from datetime import datetime
|
from datetime import datetime
|
||||||
from sqlalchemy import Column, Enum, ForeignKey, DateTime, Boolean, Integer
|
from enum import Enum as Enumeration
|
||||||
|
|
||||||
|
from sqlalchemy import Boolean, Column, DateTime, Enum, ForeignKey, Integer
|
||||||
from sqlalchemy.dialects.postgresql import JSONB
|
from sqlalchemy.dialects.postgresql import JSONB
|
||||||
|
|
||||||
from base.orm import Base
|
from base.orm import Base
|
||||||
from enum import Enum as Enumeration
|
|
||||||
|
|
||||||
|
|
||||||
class NotificationType(Enumeration):
|
class NotificationType(Enumeration):
|
||||||
|
@ -14,9 +15,9 @@ class NotificationType(Enumeration):
|
||||||
class Notification(Base):
|
class Notification(Base):
|
||||||
__tablename__ = "notification"
|
__tablename__ = "notification"
|
||||||
|
|
||||||
shout = Column(ForeignKey("shout.id"), index=True)
|
shout: Column = Column(ForeignKey("shout.id"), index=True)
|
||||||
reaction = Column(ForeignKey("reaction.id"), index=True)
|
reaction: Column = Column(ForeignKey("reaction.id"), index=True)
|
||||||
user = Column(ForeignKey("user.id"), index=True)
|
user: Column = Column(ForeignKey("user.id"), index=True)
|
||||||
createdAt = Column(DateTime, nullable=False, default=datetime.now, index=True)
|
createdAt = Column(DateTime, nullable=False, default=datetime.now, index=True)
|
||||||
seen = Column(Boolean, nullable=False, default=False, index=True)
|
seen = Column(Boolean, nullable=False, default=False, index=True)
|
||||||
type = Column(Enum(NotificationType), nullable=False)
|
type = Column(Enum(NotificationType), nullable=False)
|
||||||
|
|
49
orm/rbac.py
49
orm/rbac.py
|
@ -1,9 +1,9 @@
|
||||||
import warnings
|
import warnings
|
||||||
|
|
||||||
from sqlalchemy import String, Column, ForeignKey, UniqueConstraint, TypeDecorator
|
from sqlalchemy import Column, ForeignKey, String, TypeDecorator, UniqueConstraint
|
||||||
from sqlalchemy.orm import relationship
|
from sqlalchemy.orm import relationship
|
||||||
|
|
||||||
from base.orm import Base, REGISTRY, engine, local_session
|
from base.orm import REGISTRY, Base, local_session
|
||||||
|
|
||||||
# Role Based Access Control #
|
# Role Based Access Control #
|
||||||
|
|
||||||
|
@ -121,16 +121,23 @@ class Operation(Base):
|
||||||
|
|
||||||
class Resource(Base):
|
class Resource(Base):
|
||||||
__tablename__ = "resource"
|
__tablename__ = "resource"
|
||||||
resourceClass = Column(
|
resourceClass = Column(String, nullable=False, unique=True, comment="Resource class")
|
||||||
String, nullable=False, unique=True, comment="Resource class"
|
|
||||||
)
|
|
||||||
name = Column(String, nullable=False, unique=True, comment="Resource name")
|
name = Column(String, nullable=False, unique=True, comment="Resource name")
|
||||||
# TODO: community = Column(ForeignKey())
|
# TODO: community = Column(ForeignKey())
|
||||||
|
|
||||||
@staticmethod
|
@staticmethod
|
||||||
def init_table():
|
def init_table():
|
||||||
with local_session() as session:
|
with local_session() as session:
|
||||||
for res in ["shout", "topic", "reaction", "chat", "message", "invite", "community", "user"]:
|
for res in [
|
||||||
|
"shout",
|
||||||
|
"topic",
|
||||||
|
"reaction",
|
||||||
|
"chat",
|
||||||
|
"message",
|
||||||
|
"invite",
|
||||||
|
"community",
|
||||||
|
"user",
|
||||||
|
]:
|
||||||
r = session.query(Resource).filter(Resource.name == res).first()
|
r = session.query(Resource).filter(Resource.name == res).first()
|
||||||
if not r:
|
if not r:
|
||||||
r = Resource.create(name=res, resourceClass=res)
|
r = Resource.create(name=res, resourceClass=res)
|
||||||
|
@ -145,29 +152,27 @@ class Permission(Base):
|
||||||
{"extend_existing": True},
|
{"extend_existing": True},
|
||||||
)
|
)
|
||||||
|
|
||||||
role = Column(
|
role: Column = Column(ForeignKey("role.id", ondelete="CASCADE"), nullable=False, comment="Role")
|
||||||
ForeignKey("role.id", ondelete="CASCADE"), nullable=False, comment="Role"
|
operation: Column = Column(
|
||||||
)
|
|
||||||
operation = Column(
|
|
||||||
ForeignKey("operation.id", ondelete="CASCADE"),
|
ForeignKey("operation.id", ondelete="CASCADE"),
|
||||||
nullable=False,
|
nullable=False,
|
||||||
comment="Operation",
|
comment="Operation",
|
||||||
)
|
)
|
||||||
resource = Column(
|
resource: Column = Column(
|
||||||
ForeignKey("resource.id", ondelete="CASCADE"),
|
ForeignKey("resource.id", ondelete="CASCADE"),
|
||||||
nullable=False,
|
nullable=False,
|
||||||
comment="Resource",
|
comment="Resource",
|
||||||
)
|
)
|
||||||
|
|
||||||
|
|
||||||
if __name__ == "__main__":
|
# if __name__ == "__main__":
|
||||||
Base.metadata.create_all(engine)
|
# Base.metadata.create_all(engine)
|
||||||
ops = [
|
# ops = [
|
||||||
Permission(role=1, operation=1, resource=1),
|
# Permission(role=1, operation=1, resource=1),
|
||||||
Permission(role=1, operation=2, resource=1),
|
# Permission(role=1, operation=2, resource=1),
|
||||||
Permission(role=1, operation=3, resource=1),
|
# Permission(role=1, operation=3, resource=1),
|
||||||
Permission(role=1, operation=4, resource=1),
|
# Permission(role=1, operation=4, resource=1),
|
||||||
Permission(role=2, operation=4, resource=1),
|
# Permission(role=2, operation=4, resource=1),
|
||||||
]
|
# ]
|
||||||
global_session.add_all(ops)
|
# global_session.add_all(ops)
|
||||||
global_session.commit()
|
# global_session.commit()
|
||||||
|
|
|
@ -27,16 +27,18 @@ class ReactionKind(Enumeration):
|
||||||
class Reaction(Base):
|
class Reaction(Base):
|
||||||
__tablename__ = "reaction"
|
__tablename__ = "reaction"
|
||||||
body = Column(String, nullable=True, comment="Reaction Body")
|
body = Column(String, nullable=True, comment="Reaction Body")
|
||||||
createdAt = Column(
|
createdAt = Column(DateTime, nullable=False, default=datetime.now, comment="Created at")
|
||||||
DateTime, nullable=False, default=datetime.now, comment="Created at"
|
createdBy: Column = Column(ForeignKey("user.id"), nullable=False, index=True, comment="Sender")
|
||||||
)
|
|
||||||
createdBy = Column(ForeignKey("user.id"), nullable=False, index=True, comment="Sender")
|
|
||||||
updatedAt = Column(DateTime, nullable=True, comment="Updated at")
|
updatedAt = Column(DateTime, nullable=True, comment="Updated at")
|
||||||
updatedBy = Column(ForeignKey("user.id"), nullable=True, index=True, comment="Last Editor")
|
updatedBy: Column = Column(
|
||||||
|
ForeignKey("user.id"), nullable=True, index=True, comment="Last Editor"
|
||||||
|
)
|
||||||
deletedAt = Column(DateTime, nullable=True, comment="Deleted at")
|
deletedAt = Column(DateTime, nullable=True, comment="Deleted at")
|
||||||
deletedBy = Column(ForeignKey("user.id"), nullable=True, index=True, comment="Deleted by")
|
deletedBy: Column = Column(
|
||||||
shout = Column(ForeignKey("shout.id"), nullable=False, index=True)
|
ForeignKey("user.id"), nullable=True, index=True, comment="Deleted by"
|
||||||
replyTo = Column(
|
)
|
||||||
|
shout: Column = Column(ForeignKey("shout.id"), nullable=False, index=True)
|
||||||
|
replyTo: Column = Column(
|
||||||
ForeignKey("reaction.id"), nullable=True, comment="Reply to reaction ID"
|
ForeignKey("reaction.id"), nullable=True, comment="Reply to reaction ID"
|
||||||
)
|
)
|
||||||
range = Column(String, nullable=True, comment="Range in format <start index>:<end>")
|
range = Column(String, nullable=True, comment="Range in format <start index>:<end>")
|
||||||
|
|
45
orm/shout.py
45
orm/shout.py
|
@ -1,6 +1,6 @@
|
||||||
from datetime import datetime
|
from datetime import datetime
|
||||||
|
|
||||||
from sqlalchemy import Boolean, Column, DateTime, ForeignKey, Integer, String, JSON
|
from sqlalchemy import JSON, Boolean, Column, DateTime, ForeignKey, Integer, String
|
||||||
from sqlalchemy.orm import column_property, relationship
|
from sqlalchemy.orm import column_property, relationship
|
||||||
|
|
||||||
from base.orm import Base, local_session
|
from base.orm import Base, local_session
|
||||||
|
@ -12,31 +12,29 @@ from orm.user import User
|
||||||
class ShoutTopic(Base):
|
class ShoutTopic(Base):
|
||||||
__tablename__ = "shout_topic"
|
__tablename__ = "shout_topic"
|
||||||
|
|
||||||
id = None # type: ignore
|
id = None
|
||||||
shout = Column(ForeignKey("shout.id"), primary_key=True, index=True)
|
shout: Column = Column(ForeignKey("shout.id"), primary_key=True, index=True)
|
||||||
topic = Column(ForeignKey("topic.id"), primary_key=True, index=True)
|
topic: Column = Column(ForeignKey("topic.id"), primary_key=True, index=True)
|
||||||
|
|
||||||
|
|
||||||
class ShoutReactionsFollower(Base):
|
class ShoutReactionsFollower(Base):
|
||||||
__tablename__ = "shout_reactions_followers"
|
__tablename__ = "shout_reactions_followers"
|
||||||
|
|
||||||
id = None # type: ignore
|
id = None
|
||||||
follower = Column(ForeignKey("user.id"), primary_key=True, index=True)
|
follower: Column = Column(ForeignKey("user.id"), primary_key=True, index=True)
|
||||||
shout = Column(ForeignKey("shout.id"), primary_key=True, index=True)
|
shout: Column = Column(ForeignKey("shout.id"), primary_key=True, index=True)
|
||||||
auto = Column(Boolean, nullable=False, default=False)
|
auto = Column(Boolean, nullable=False, default=False)
|
||||||
createdAt = Column(
|
createdAt = Column(DateTime, nullable=False, default=datetime.now, comment="Created at")
|
||||||
DateTime, nullable=False, default=datetime.now, comment="Created at"
|
|
||||||
)
|
|
||||||
deletedAt = Column(DateTime, nullable=True)
|
deletedAt = Column(DateTime, nullable=True)
|
||||||
|
|
||||||
|
|
||||||
class ShoutAuthor(Base):
|
class ShoutAuthor(Base):
|
||||||
__tablename__ = "shout_author"
|
__tablename__ = "shout_author"
|
||||||
|
|
||||||
id = None # type: ignore
|
id = None
|
||||||
shout = Column(ForeignKey("shout.id"), primary_key=True, index=True)
|
shout: Column = Column(ForeignKey("shout.id"), primary_key=True, index=True)
|
||||||
user = Column(ForeignKey("user.id"), primary_key=True, index=True)
|
user: Column = Column(ForeignKey("user.id"), primary_key=True, index=True)
|
||||||
caption = Column(String, nullable=True, default="")
|
caption: Column = Column(String, nullable=True, default="")
|
||||||
|
|
||||||
|
|
||||||
class Shout(Base):
|
class Shout(Base):
|
||||||
|
@ -48,8 +46,8 @@ class Shout(Base):
|
||||||
publishedAt = Column(DateTime, nullable=True)
|
publishedAt = Column(DateTime, nullable=True)
|
||||||
deletedAt = Column(DateTime, nullable=True)
|
deletedAt = Column(DateTime, nullable=True)
|
||||||
|
|
||||||
createdBy = Column(ForeignKey("user.id"), comment="Created By")
|
createdBy: Column = Column(ForeignKey("user.id"), comment="Created By")
|
||||||
deletedBy = Column(ForeignKey("user.id"), nullable=True)
|
deletedBy: Column = Column(ForeignKey("user.id"), nullable=True)
|
||||||
|
|
||||||
slug = Column(String, unique=True)
|
slug = Column(String, unique=True)
|
||||||
cover = Column(String, nullable=True, comment="Cover image url")
|
cover = Column(String, nullable=True, comment="Cover image url")
|
||||||
|
@ -71,11 +69,11 @@ class Shout(Base):
|
||||||
reactions = relationship(lambda: Reaction)
|
reactions = relationship(lambda: Reaction)
|
||||||
|
|
||||||
# TODO: these field should be used or modified
|
# TODO: these field should be used or modified
|
||||||
community = Column(ForeignKey("community.id"), default=1)
|
community: Column = Column(ForeignKey("community.id"), default=1)
|
||||||
lang = Column(String, nullable=False, default='ru', comment="Language")
|
lang = Column(String, nullable=False, default="ru", comment="Language")
|
||||||
mainTopic = Column(ForeignKey("topic.slug"), nullable=True)
|
mainTopic: Column = Column(ForeignKey("topic.slug"), nullable=True)
|
||||||
visibility = Column(String, nullable=True) # owner authors community public
|
visibility = Column(String, nullable=True) # owner authors community public
|
||||||
versionOf = Column(ForeignKey("shout.id"), nullable=True)
|
versionOf: Column = Column(ForeignKey("shout.id"), nullable=True)
|
||||||
oid = Column(String, nullable=True)
|
oid = Column(String, nullable=True)
|
||||||
|
|
||||||
@staticmethod
|
@staticmethod
|
||||||
|
@ -83,12 +81,7 @@ class Shout(Base):
|
||||||
with local_session() as session:
|
with local_session() as session:
|
||||||
s = session.query(Shout).first()
|
s = session.query(Shout).first()
|
||||||
if not s:
|
if not s:
|
||||||
entry = {
|
entry = {"slug": "genesis-block", "body": "", "title": "Ничего", "lang": "ru"}
|
||||||
"slug": "genesis-block",
|
|
||||||
"body": "",
|
|
||||||
"title": "Ничего",
|
|
||||||
"lang": "ru"
|
|
||||||
}
|
|
||||||
s = Shout.create(**entry)
|
s = Shout.create(**entry)
|
||||||
session.add(s)
|
session.add(s)
|
||||||
session.commit()
|
session.commit()
|
||||||
|
|
14
orm/topic.py
14
orm/topic.py
|
@ -8,12 +8,10 @@ from base.orm import Base
|
||||||
class TopicFollower(Base):
|
class TopicFollower(Base):
|
||||||
__tablename__ = "topic_followers"
|
__tablename__ = "topic_followers"
|
||||||
|
|
||||||
id = None # type: ignore
|
id = None
|
||||||
follower = Column(ForeignKey("user.id"), primary_key=True, index=True)
|
follower: Column = Column(ForeignKey("user.id"), primary_key=True, index=True)
|
||||||
topic = Column(ForeignKey("topic.id"), primary_key=True, index=True)
|
topic: Column = Column(ForeignKey("topic.id"), primary_key=True, index=True)
|
||||||
createdAt = Column(
|
createdAt = Column(DateTime, nullable=False, default=datetime.now, comment="Created at")
|
||||||
DateTime, nullable=False, default=datetime.now, comment="Created at"
|
|
||||||
)
|
|
||||||
auto = Column(Boolean, nullable=False, default=False)
|
auto = Column(Boolean, nullable=False, default=False)
|
||||||
|
|
||||||
|
|
||||||
|
@ -24,7 +22,5 @@ class Topic(Base):
|
||||||
title = Column(String, nullable=False, comment="Title")
|
title = Column(String, nullable=False, comment="Title")
|
||||||
body = Column(String, nullable=True, comment="Body")
|
body = Column(String, nullable=True, comment="Body")
|
||||||
pic = Column(String, nullable=True, comment="Picture")
|
pic = Column(String, nullable=True, comment="Picture")
|
||||||
community = Column(
|
community: Column = Column(ForeignKey("community.id"), default=1, comment="Community")
|
||||||
ForeignKey("community.id"), default=1, comment="Community"
|
|
||||||
)
|
|
||||||
oid = Column(String, nullable=True, comment="Old ID")
|
oid = Column(String, nullable=True, comment="Old ID")
|
||||||
|
|
31
orm/user.py
31
orm/user.py
|
@ -3,6 +3,7 @@ from datetime import datetime
|
||||||
from sqlalchemy import JSON as JSONType
|
from sqlalchemy import JSON as JSONType
|
||||||
from sqlalchemy import Boolean, Column, DateTime, ForeignKey, Integer, String
|
from sqlalchemy import Boolean, Column, DateTime, ForeignKey, Integer, String
|
||||||
from sqlalchemy.orm import relationship
|
from sqlalchemy.orm import relationship
|
||||||
|
|
||||||
from base.orm import Base, local_session
|
from base.orm import Base, local_session
|
||||||
from orm.rbac import Role
|
from orm.rbac import Role
|
||||||
|
|
||||||
|
@ -10,10 +11,10 @@ from orm.rbac import Role
|
||||||
class UserRating(Base):
|
class UserRating(Base):
|
||||||
__tablename__ = "user_rating"
|
__tablename__ = "user_rating"
|
||||||
|
|
||||||
id = None # type: ignore
|
id = None
|
||||||
rater = Column(ForeignKey("user.id"), primary_key=True, index=True)
|
rater: Column = Column(ForeignKey("user.id"), primary_key=True, index=True)
|
||||||
user = Column(ForeignKey("user.id"), primary_key=True, index=True)
|
user: Column = Column(ForeignKey("user.id"), primary_key=True, index=True)
|
||||||
value = Column(Integer)
|
value: Column = Column(Integer)
|
||||||
|
|
||||||
@staticmethod
|
@staticmethod
|
||||||
def init_table():
|
def init_table():
|
||||||
|
@ -23,7 +24,7 @@ class UserRating(Base):
|
||||||
class UserRole(Base):
|
class UserRole(Base):
|
||||||
__tablename__ = "user_role"
|
__tablename__ = "user_role"
|
||||||
|
|
||||||
id = None # type: ignore
|
id = None
|
||||||
user = Column(ForeignKey("user.id"), primary_key=True, index=True)
|
user = Column(ForeignKey("user.id"), primary_key=True, index=True)
|
||||||
role = Column(ForeignKey("role.id"), primary_key=True, index=True)
|
role = Column(ForeignKey("role.id"), primary_key=True, index=True)
|
||||||
|
|
||||||
|
@ -31,12 +32,10 @@ class UserRole(Base):
|
||||||
class AuthorFollower(Base):
|
class AuthorFollower(Base):
|
||||||
__tablename__ = "author_follower"
|
__tablename__ = "author_follower"
|
||||||
|
|
||||||
id = None # type: ignore
|
id = None
|
||||||
follower = Column(ForeignKey("user.id"), primary_key=True, index=True)
|
follower: Column = Column(ForeignKey("user.id"), primary_key=True, index=True)
|
||||||
author = Column(ForeignKey("user.id"), primary_key=True, index=True)
|
author: Column = Column(ForeignKey("user.id"), primary_key=True, index=True)
|
||||||
createdAt = Column(
|
createdAt = Column(DateTime, nullable=False, default=datetime.now, comment="Created at")
|
||||||
DateTime, nullable=False, default=datetime.now, comment="Created at"
|
|
||||||
)
|
|
||||||
auto = Column(Boolean, nullable=False, default=False)
|
auto = Column(Boolean, nullable=False, default=False)
|
||||||
|
|
||||||
|
|
||||||
|
@ -54,12 +53,8 @@ class User(Base):
|
||||||
slug = Column(String, unique=True, comment="User's slug")
|
slug = Column(String, unique=True, comment="User's slug")
|
||||||
muted = Column(Boolean, default=False)
|
muted = Column(Boolean, default=False)
|
||||||
emailConfirmed = Column(Boolean, default=False)
|
emailConfirmed = Column(Boolean, default=False)
|
||||||
createdAt = Column(
|
createdAt = Column(DateTime, nullable=False, default=datetime.now, comment="Created at")
|
||||||
DateTime, nullable=False, default=datetime.now, comment="Created at"
|
lastSeen = Column(DateTime, nullable=False, default=datetime.now, comment="Was online at")
|
||||||
)
|
|
||||||
lastSeen = Column(
|
|
||||||
DateTime, nullable=False, default=datetime.now, comment="Was online at"
|
|
||||||
)
|
|
||||||
deletedAt = Column(DateTime, nullable=True, comment="Deleted at")
|
deletedAt = Column(DateTime, nullable=True, comment="Deleted at")
|
||||||
links = Column(JSONType, nullable=True, comment="Links")
|
links = Column(JSONType, nullable=True, comment="Links")
|
||||||
oauth = Column(String, nullable=True)
|
oauth = Column(String, nullable=True)
|
||||||
|
@ -103,4 +98,4 @@ class User(Base):
|
||||||
|
|
||||||
|
|
||||||
# if __name__ == "__main__":
|
# if __name__ == "__main__":
|
||||||
# print(User.get_permission(user_id=1)) # type: ignore
|
# print(User.get_permission(user_id=1))
|
||||||
|
|
2
pyproject.toml
Normal file
2
pyproject.toml
Normal file
|
@ -0,0 +1,2 @@
|
||||||
|
[tool.black]
|
||||||
|
line-length = 100
|
|
@ -1,4 +1,8 @@
|
||||||
isort
|
black==23.10.1
|
||||||
brunette
|
flake8==6.1.0
|
||||||
flake8
|
gql_schema_codegen==1.0.1
|
||||||
mypy
|
isort==5.12.0
|
||||||
|
mypy==1.6.1
|
||||||
|
pre-commit==3.5.0
|
||||||
|
pymongo-stubs==0.2.0
|
||||||
|
sqlalchemy-stubs==0.4
|
||||||
|
|
|
@ -1,40 +1,37 @@
|
||||||
python-frontmatter~=1.0.0
|
|
||||||
aioredis~=2.0.1
|
|
||||||
aiohttp
|
aiohttp
|
||||||
|
aioredis~=2.0.1
|
||||||
|
alembic==1.11.3
|
||||||
ariadne>=0.17.0
|
ariadne>=0.17.0
|
||||||
PyYAML>=5.4
|
|
||||||
pyjwt>=2.6.0
|
|
||||||
starlette~=0.23.1
|
|
||||||
sqlalchemy>=1.4.41
|
|
||||||
graphql-core>=3.0.3
|
|
||||||
gql~=3.4.0
|
|
||||||
uvicorn>=0.18.3
|
|
||||||
pydantic>=1.10.2
|
|
||||||
passlib~=1.7.4
|
|
||||||
authlib>=1.1.0
|
|
||||||
httpx>=0.23.0
|
|
||||||
psycopg2-binary
|
|
||||||
transliterate~=1.10.2
|
|
||||||
requests~=2.28.1
|
|
||||||
bcrypt>=4.0.0
|
|
||||||
bson~=0.5.10
|
|
||||||
flake8
|
|
||||||
DateTime~=4.7
|
|
||||||
asyncio~=3.4.3
|
asyncio~=3.4.3
|
||||||
python-dateutil~=2.8.2
|
authlib>=1.1.0
|
||||||
|
bcrypt>=4.0.0
|
||||||
beautifulsoup4~=4.11.1
|
beautifulsoup4~=4.11.1
|
||||||
lxml
|
|
||||||
sentry-sdk>=1.14.0
|
|
||||||
# sse_starlette
|
|
||||||
graphql-ws
|
|
||||||
nltk~=3.8.1
|
|
||||||
pymystem3~=0.2.0
|
|
||||||
transformers~=4.28.1
|
|
||||||
boto3~=1.28.2
|
boto3~=1.28.2
|
||||||
botocore~=1.31.2
|
botocore~=1.31.2
|
||||||
python-multipart~=0.0.6
|
bson~=0.5.10
|
||||||
alembic==1.11.3
|
DateTime~=4.7
|
||||||
|
gql~=3.4.0
|
||||||
|
graphql-core>=3.0.3
|
||||||
|
httpx>=0.23.0
|
||||||
|
itsdangerous
|
||||||
|
lxml
|
||||||
Mako==1.2.4
|
Mako==1.2.4
|
||||||
MarkupSafe==2.1.3
|
MarkupSafe==2.1.3
|
||||||
|
nltk~=3.8.1
|
||||||
|
passlib~=1.7.4
|
||||||
|
psycopg2-binary
|
||||||
|
pydantic>=1.10.2
|
||||||
|
pyjwt>=2.6.0
|
||||||
|
pymystem3~=0.2.0
|
||||||
|
python-dateutil~=2.8.2
|
||||||
|
python-frontmatter~=1.0.0
|
||||||
|
python-multipart~=0.0.6
|
||||||
|
PyYAML>=5.4
|
||||||
|
requests~=2.28.1
|
||||||
|
sentry-sdk>=1.14.0
|
||||||
|
sqlalchemy>=1.4.41
|
||||||
sse-starlette==1.6.5
|
sse-starlette==1.6.5
|
||||||
itsdangerous
|
starlette~=0.23.1
|
||||||
|
transformers~=4.28.1
|
||||||
|
transliterate~=1.10.2
|
||||||
|
uvicorn>=0.18.3
|
||||||
|
|
|
@ -53,4 +53,3 @@ echo "Start migration"
|
||||||
python3 server.py migrate
|
python3 server.py migrate
|
||||||
if [ $? -ne 0 ]; then { echo "Migration failed, aborting." ; exit 1; } fi
|
if [ $? -ne 0 ]; then { echo "Migration failed, aborting." ; exit 1; } fi
|
||||||
echo 'Done!'
|
echo 'Done!'
|
||||||
|
|
||||||
|
|
|
@ -1,67 +1,46 @@
|
||||||
|
# flake8: noqa
|
||||||
|
|
||||||
from resolvers.auth import (
|
from resolvers.auth import (
|
||||||
login,
|
|
||||||
sign_out,
|
|
||||||
is_email_used,
|
|
||||||
register_by_email,
|
|
||||||
confirm_email,
|
|
||||||
auth_send_link,
|
auth_send_link,
|
||||||
|
confirm_email,
|
||||||
get_current_user,
|
get_current_user,
|
||||||
|
is_email_used,
|
||||||
|
login,
|
||||||
|
register_by_email,
|
||||||
|
sign_out,
|
||||||
)
|
)
|
||||||
|
|
||||||
from resolvers.create.migrate import markdown_body
|
|
||||||
from resolvers.create.editor import create_shout, delete_shout, update_shout
|
from resolvers.create.editor import create_shout, delete_shout, update_shout
|
||||||
|
from resolvers.inbox.chats import create_chat, delete_chat, update_chat
|
||||||
from resolvers.zine.profile import (
|
from resolvers.inbox.load import load_chats, load_messages_by, load_recipients
|
||||||
load_authors_by,
|
|
||||||
rate_user,
|
|
||||||
update_profile,
|
|
||||||
get_authors_all
|
|
||||||
)
|
|
||||||
|
|
||||||
from resolvers.zine.reactions import (
|
|
||||||
create_reaction,
|
|
||||||
delete_reaction,
|
|
||||||
update_reaction,
|
|
||||||
reactions_unfollow,
|
|
||||||
reactions_follow,
|
|
||||||
load_reactions_by
|
|
||||||
)
|
|
||||||
from resolvers.zine.topics import (
|
|
||||||
topic_follow,
|
|
||||||
topic_unfollow,
|
|
||||||
topics_by_author,
|
|
||||||
topics_by_community,
|
|
||||||
topics_all,
|
|
||||||
get_topic
|
|
||||||
)
|
|
||||||
|
|
||||||
from resolvers.zine.following import (
|
|
||||||
follow,
|
|
||||||
unfollow
|
|
||||||
)
|
|
||||||
|
|
||||||
from resolvers.zine.load import (
|
|
||||||
load_shout,
|
|
||||||
load_shouts_by
|
|
||||||
)
|
|
||||||
|
|
||||||
from resolvers.inbox.chats import (
|
|
||||||
create_chat,
|
|
||||||
delete_chat,
|
|
||||||
update_chat
|
|
||||||
|
|
||||||
)
|
|
||||||
from resolvers.inbox.messages import (
|
from resolvers.inbox.messages import (
|
||||||
create_message,
|
create_message,
|
||||||
delete_message,
|
delete_message,
|
||||||
|
mark_as_read,
|
||||||
update_message,
|
update_message,
|
||||||
mark_as_read
|
|
||||||
)
|
|
||||||
from resolvers.inbox.load import (
|
|
||||||
load_chats,
|
|
||||||
load_messages_by,
|
|
||||||
load_recipients
|
|
||||||
)
|
)
|
||||||
from resolvers.inbox.search import search_recipients
|
from resolvers.inbox.search import search_recipients
|
||||||
|
|
||||||
from resolvers.notifications import load_notifications
|
from resolvers.notifications import load_notifications
|
||||||
|
from resolvers.zine.following import follow, unfollow
|
||||||
|
from resolvers.zine.load import load_shout, load_shouts_by
|
||||||
|
from resolvers.zine.profile import (
|
||||||
|
get_authors_all,
|
||||||
|
load_authors_by,
|
||||||
|
rate_user,
|
||||||
|
update_profile,
|
||||||
|
)
|
||||||
|
from resolvers.zine.reactions import (
|
||||||
|
create_reaction,
|
||||||
|
delete_reaction,
|
||||||
|
load_reactions_by,
|
||||||
|
reactions_follow,
|
||||||
|
reactions_unfollow,
|
||||||
|
update_reaction,
|
||||||
|
)
|
||||||
|
from resolvers.zine.topics import (
|
||||||
|
get_topic,
|
||||||
|
topic_follow,
|
||||||
|
topic_unfollow,
|
||||||
|
topics_all,
|
||||||
|
topics_by_author,
|
||||||
|
topics_by_community,
|
||||||
|
)
|
||||||
|
|
|
@ -1,24 +1,30 @@
|
||||||
# -*- coding: utf-8 -*-
|
# -*- coding: utf-8 -*-
|
||||||
|
|
||||||
|
import re
|
||||||
from datetime import datetime, timezone
|
from datetime import datetime, timezone
|
||||||
from urllib.parse import quote_plus
|
from urllib.parse import quote_plus
|
||||||
|
|
||||||
from graphql.type import GraphQLResolveInfo
|
from graphql.type import GraphQLResolveInfo
|
||||||
from starlette.responses import RedirectResponse
|
from starlette.responses import RedirectResponse
|
||||||
from transliterate import translit
|
from transliterate import translit
|
||||||
import re
|
|
||||||
from auth.authenticate import login_required
|
from auth.authenticate import login_required
|
||||||
from auth.credentials import AuthCredentials
|
from auth.credentials import AuthCredentials
|
||||||
from auth.email import send_auth_email
|
from auth.email import send_auth_email
|
||||||
from auth.identity import Identity, Password
|
from auth.identity import Identity, Password
|
||||||
from auth.jwtcodec import JWTCodec
|
from auth.jwtcodec import JWTCodec
|
||||||
from auth.tokenstorage import TokenStorage
|
from auth.tokenstorage import TokenStorage
|
||||||
from base.exceptions import (BaseHttpException, InvalidPassword, InvalidToken,
|
from base.exceptions import (
|
||||||
ObjectNotExist, Unauthorized)
|
BaseHttpException,
|
||||||
|
InvalidPassword,
|
||||||
|
InvalidToken,
|
||||||
|
ObjectNotExist,
|
||||||
|
Unauthorized,
|
||||||
|
)
|
||||||
from base.orm import local_session
|
from base.orm import local_session
|
||||||
from base.resolvers import mutation, query
|
from base.resolvers import mutation, query
|
||||||
from orm import Role, User
|
from orm import Role, User
|
||||||
from settings import SESSION_TOKEN_HEADER, FRONTEND_URL
|
from settings import FRONTEND_URL, SESSION_TOKEN_HEADER
|
||||||
|
|
||||||
|
|
||||||
@mutation.field("getSession")
|
@mutation.field("getSession")
|
||||||
|
@ -32,17 +38,14 @@ async def get_current_user(_, info):
|
||||||
user.lastSeen = datetime.now(tz=timezone.utc)
|
user.lastSeen = datetime.now(tz=timezone.utc)
|
||||||
session.commit()
|
session.commit()
|
||||||
|
|
||||||
return {
|
return {"token": token, "user": user}
|
||||||
"token": token,
|
|
||||||
"user": user
|
|
||||||
}
|
|
||||||
|
|
||||||
|
|
||||||
@mutation.field("confirmEmail")
|
@mutation.field("confirmEmail")
|
||||||
async def confirm_email(_, info, token):
|
async def confirm_email(_, info, token):
|
||||||
"""confirm owning email address"""
|
"""confirm owning email address"""
|
||||||
try:
|
try:
|
||||||
print('[resolvers.auth] confirm email by token')
|
print("[resolvers.auth] confirm email by token")
|
||||||
payload = JWTCodec.decode(token)
|
payload = JWTCodec.decode(token)
|
||||||
user_id = payload.user_id
|
user_id = payload.user_id
|
||||||
await TokenStorage.get(f"{user_id}-{payload.username}-{token}")
|
await TokenStorage.get(f"{user_id}-{payload.username}-{token}")
|
||||||
|
@ -53,10 +56,7 @@ async def confirm_email(_, info, token):
|
||||||
user.lastSeen = datetime.now(tz=timezone.utc)
|
user.lastSeen = datetime.now(tz=timezone.utc)
|
||||||
session.add(user)
|
session.add(user)
|
||||||
session.commit()
|
session.commit()
|
||||||
return {
|
return {"token": session_token, "user": user}
|
||||||
"token": session_token,
|
|
||||||
"user": user
|
|
||||||
}
|
|
||||||
except InvalidToken as e:
|
except InvalidToken as e:
|
||||||
raise InvalidToken(e.message)
|
raise InvalidToken(e.message)
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
|
@ -68,9 +68,9 @@ async def confirm_email_handler(request):
|
||||||
token = request.path_params["token"] # one time
|
token = request.path_params["token"] # one time
|
||||||
request.session["token"] = token
|
request.session["token"] = token
|
||||||
res = await confirm_email(None, {}, token)
|
res = await confirm_email(None, {}, token)
|
||||||
print('[resolvers.auth] confirm_email request: %r' % request)
|
print("[resolvers.auth] confirm_email request: %r" % request)
|
||||||
if "error" in res:
|
if "error" in res:
|
||||||
raise BaseHttpException(res['error'])
|
raise BaseHttpException(res["error"])
|
||||||
else:
|
else:
|
||||||
response = RedirectResponse(url=FRONTEND_URL)
|
response = RedirectResponse(url=FRONTEND_URL)
|
||||||
response.set_cookie("token", res["token"]) # session token
|
response.set_cookie("token", res["token"]) # session token
|
||||||
|
@ -87,22 +87,22 @@ def create_user(user_dict):
|
||||||
|
|
||||||
|
|
||||||
def generate_unique_slug(src):
|
def generate_unique_slug(src):
|
||||||
print('[resolvers.auth] generating slug from: ' + src)
|
print("[resolvers.auth] generating slug from: " + src)
|
||||||
slug = translit(src, "ru", reversed=True).replace(".", "-").lower()
|
slug = translit(src, "ru", reversed=True).replace(".", "-").lower()
|
||||||
slug = re.sub('[^0-9a-zA-Z]+', '-', slug)
|
slug = re.sub("[^0-9a-zA-Z]+", "-", slug)
|
||||||
if slug != src:
|
if slug != src:
|
||||||
print('[resolvers.auth] translited name: ' + slug)
|
print("[resolvers.auth] translited name: " + slug)
|
||||||
c = 1
|
c = 1
|
||||||
with local_session() as session:
|
with local_session() as session:
|
||||||
user = session.query(User).where(User.slug == slug).first()
|
user = session.query(User).where(User.slug == slug).first()
|
||||||
while user:
|
while user:
|
||||||
user = session.query(User).where(User.slug == slug).first()
|
user = session.query(User).where(User.slug == slug).first()
|
||||||
slug = slug + '-' + str(c)
|
slug = slug + "-" + str(c)
|
||||||
c += 1
|
c += 1
|
||||||
if not user:
|
if not user:
|
||||||
unique_slug = slug
|
unique_slug = slug
|
||||||
print('[resolvers.auth] ' + unique_slug)
|
print("[resolvers.auth] " + unique_slug)
|
||||||
return quote_plus(unique_slug.replace('\'', '')).replace('+', '-')
|
return quote_plus(unique_slug.replace("'", "")).replace("+", "-")
|
||||||
|
|
||||||
|
|
||||||
@mutation.field("registerUser")
|
@mutation.field("registerUser")
|
||||||
|
@ -117,12 +117,12 @@ async def register_by_email(_, _info, email: str, password: str = "", name: str
|
||||||
slug = generate_unique_slug(name)
|
slug = generate_unique_slug(name)
|
||||||
user = session.query(User).where(User.slug == slug).first()
|
user = session.query(User).where(User.slug == slug).first()
|
||||||
if user:
|
if user:
|
||||||
slug = generate_unique_slug(email.split('@')[0])
|
slug = generate_unique_slug(email.split("@")[0])
|
||||||
user_dict = {
|
user_dict = {
|
||||||
"email": email,
|
"email": email,
|
||||||
"username": email, # will be used to store phone number or some messenger network id
|
"username": email, # will be used to store phone number or some messenger network id
|
||||||
"name": name,
|
"name": name,
|
||||||
"slug": slug
|
"slug": slug,
|
||||||
}
|
}
|
||||||
if password:
|
if password:
|
||||||
user_dict["password"] = Password.encode(password)
|
user_dict["password"] = Password.encode(password)
|
||||||
|
@ -172,10 +172,7 @@ async def login(_, info, email: str, password: str = "", lang: str = "ru"):
|
||||||
user = Identity.password(orm_user, password)
|
user = Identity.password(orm_user, password)
|
||||||
session_token = await TokenStorage.create_session(user)
|
session_token = await TokenStorage.create_session(user)
|
||||||
print(f"[auth] user {email} authorized")
|
print(f"[auth] user {email} authorized")
|
||||||
return {
|
return {"token": session_token, "user": user}
|
||||||
"token": session_token,
|
|
||||||
"user": user
|
|
||||||
}
|
|
||||||
except InvalidPassword:
|
except InvalidPassword:
|
||||||
print(f"[auth] {email}: invalid password")
|
print(f"[auth] {email}: invalid password")
|
||||||
raise InvalidPassword("invalid password") # contains webserver status
|
raise InvalidPassword("invalid password") # contains webserver status
|
||||||
|
|
|
@ -18,21 +18,23 @@ async def create_shout(_, info, inp):
|
||||||
auth: AuthCredentials = info.context["request"].auth
|
auth: AuthCredentials = info.context["request"].auth
|
||||||
|
|
||||||
with local_session() as session:
|
with local_session() as session:
|
||||||
topics = session.query(Topic).filter(Topic.slug.in_(inp.get('topics', []))).all()
|
topics = session.query(Topic).filter(Topic.slug.in_(inp.get("topics", []))).all()
|
||||||
|
|
||||||
new_shout = Shout.create(**{
|
new_shout = Shout.create(
|
||||||
"title": inp.get("title"),
|
**{
|
||||||
"subtitle": inp.get('subtitle'),
|
"title": inp.get("title"),
|
||||||
"lead": inp.get('lead'),
|
"subtitle": inp.get("subtitle"),
|
||||||
"description": inp.get('description'),
|
"lead": inp.get("lead"),
|
||||||
"body": inp.get("body", ''),
|
"description": inp.get("description"),
|
||||||
"layout": inp.get("layout"),
|
"body": inp.get("body", ""),
|
||||||
"authors": inp.get("authors", []),
|
"layout": inp.get("layout"),
|
||||||
"slug": inp.get("slug"),
|
"authors": inp.get("authors", []),
|
||||||
"mainTopic": inp.get("mainTopic"),
|
"slug": inp.get("slug"),
|
||||||
"visibility": "owner",
|
"mainTopic": inp.get("mainTopic"),
|
||||||
"createdBy": auth.user_id
|
"visibility": "owner",
|
||||||
})
|
"createdBy": auth.user_id,
|
||||||
|
}
|
||||||
|
)
|
||||||
|
|
||||||
for topic in topics:
|
for topic in topics:
|
||||||
t = ShoutTopic.create(topic=topic.id, shout=new_shout.id)
|
t = ShoutTopic.create(topic=topic.id, shout=new_shout.id)
|
||||||
|
@ -60,14 +62,19 @@ async def create_shout(_, info, inp):
|
||||||
|
|
||||||
@mutation.field("updateShout")
|
@mutation.field("updateShout")
|
||||||
@login_required
|
@login_required
|
||||||
async def update_shout(_, info, shout_id, shout_input=None, publish=False):
|
async def update_shout(_, info, shout_id, shout_input=None, publish=False): # noqa: C901
|
||||||
auth: AuthCredentials = info.context["request"].auth
|
auth: AuthCredentials = info.context["request"].auth
|
||||||
|
|
||||||
with local_session() as session:
|
with local_session() as session:
|
||||||
shout = session.query(Shout).options(
|
shout = (
|
||||||
joinedload(Shout.authors),
|
session.query(Shout)
|
||||||
joinedload(Shout.topics),
|
.options(
|
||||||
).filter(Shout.id == shout_id).first()
|
joinedload(Shout.authors),
|
||||||
|
joinedload(Shout.topics),
|
||||||
|
)
|
||||||
|
.filter(Shout.id == shout_id)
|
||||||
|
.first()
|
||||||
|
)
|
||||||
|
|
||||||
if not shout:
|
if not shout:
|
||||||
return {"error": "shout not found"}
|
return {"error": "shout not found"}
|
||||||
|
@ -94,25 +101,34 @@ async def update_shout(_, info, shout_id, shout_input=None, publish=False):
|
||||||
session.commit()
|
session.commit()
|
||||||
|
|
||||||
for new_topic_to_link in new_topics_to_link:
|
for new_topic_to_link in new_topics_to_link:
|
||||||
created_unlinked_topic = ShoutTopic.create(shout=shout.id, topic=new_topic_to_link.id)
|
created_unlinked_topic = ShoutTopic.create(
|
||||||
|
shout=shout.id, topic=new_topic_to_link.id
|
||||||
|
)
|
||||||
session.add(created_unlinked_topic)
|
session.add(created_unlinked_topic)
|
||||||
|
|
||||||
existing_topics_input = [topic_input for topic_input in topics_input if topic_input.get("id", 0) > 0]
|
existing_topics_input = [
|
||||||
existing_topic_to_link_ids = [existing_topic_input["id"] for existing_topic_input in existing_topics_input
|
topic_input for topic_input in topics_input if topic_input.get("id", 0) > 0
|
||||||
if existing_topic_input["id"] not in [topic.id for topic in shout.topics]]
|
]
|
||||||
|
existing_topic_to_link_ids = [
|
||||||
|
existing_topic_input["id"]
|
||||||
|
for existing_topic_input in existing_topics_input
|
||||||
|
if existing_topic_input["id"] not in [topic.id for topic in shout.topics]
|
||||||
|
]
|
||||||
|
|
||||||
for existing_topic_to_link_id in existing_topic_to_link_ids:
|
for existing_topic_to_link_id in existing_topic_to_link_ids:
|
||||||
created_unlinked_topic = ShoutTopic.create(shout=shout.id, topic=existing_topic_to_link_id)
|
created_unlinked_topic = ShoutTopic.create(
|
||||||
|
shout=shout.id, topic=existing_topic_to_link_id
|
||||||
|
)
|
||||||
session.add(created_unlinked_topic)
|
session.add(created_unlinked_topic)
|
||||||
|
|
||||||
topic_to_unlink_ids = [topic.id for topic in shout.topics
|
topic_to_unlink_ids = [
|
||||||
if topic.id not in [topic_input["id"] for topic_input in existing_topics_input]]
|
topic.id
|
||||||
|
for topic in shout.topics
|
||||||
|
if topic.id not in [topic_input["id"] for topic_input in existing_topics_input]
|
||||||
|
]
|
||||||
|
|
||||||
shout_topics_to_remove = session.query(ShoutTopic).filter(
|
shout_topics_to_remove = session.query(ShoutTopic).filter(
|
||||||
and_(
|
and_(ShoutTopic.shout == shout.id, ShoutTopic.topic.in_(topic_to_unlink_ids))
|
||||||
ShoutTopic.shout == shout.id,
|
|
||||||
ShoutTopic.topic.in_(topic_to_unlink_ids)
|
|
||||||
)
|
|
||||||
)
|
)
|
||||||
|
|
||||||
for shout_topic_to_remove in shout_topics_to_remove:
|
for shout_topic_to_remove in shout_topics_to_remove:
|
||||||
|
@ -120,13 +136,13 @@ async def update_shout(_, info, shout_id, shout_input=None, publish=False):
|
||||||
|
|
||||||
shout_input["mainTopic"] = shout_input["mainTopic"]["slug"]
|
shout_input["mainTopic"] = shout_input["mainTopic"]["slug"]
|
||||||
|
|
||||||
if shout_input["mainTopic"] == '':
|
if shout_input["mainTopic"] == "":
|
||||||
del shout_input["mainTopic"]
|
del shout_input["mainTopic"]
|
||||||
|
|
||||||
shout.update(shout_input)
|
shout.update(shout_input)
|
||||||
updated = True
|
updated = True
|
||||||
|
|
||||||
if publish and shout.visibility == 'owner':
|
if publish and shout.visibility == "owner":
|
||||||
shout.visibility = "community"
|
shout.visibility = "community"
|
||||||
shout.publishedAt = datetime.now(tz=timezone.utc)
|
shout.publishedAt = datetime.now(tz=timezone.utc)
|
||||||
updated = True
|
updated = True
|
||||||
|
|
|
@ -1,11 +0,0 @@
|
||||||
|
|
||||||
from base.resolvers import query
|
|
||||||
from resolvers.auth import login_required
|
|
||||||
from migration.extract import extract_md
|
|
||||||
|
|
||||||
|
|
||||||
@login_required
|
|
||||||
@query.field("markdownBody")
|
|
||||||
def markdown_body(_, info, body: str):
|
|
||||||
body = extract_md(body)
|
|
||||||
return body
|
|
|
@ -24,27 +24,24 @@ async def update_chat(_, info, chat_new: Chat):
|
||||||
chat_id = chat_new["id"]
|
chat_id = chat_new["id"]
|
||||||
chat = await redis.execute("GET", f"chats/{chat_id}")
|
chat = await redis.execute("GET", f"chats/{chat_id}")
|
||||||
if not chat:
|
if not chat:
|
||||||
return {
|
return {"error": "chat not exist"}
|
||||||
"error": "chat not exist"
|
|
||||||
}
|
|
||||||
chat = dict(json.loads(chat))
|
chat = dict(json.loads(chat))
|
||||||
|
|
||||||
# TODO
|
# TODO
|
||||||
if auth.user_id in chat["admins"]:
|
if auth.user_id in chat["admins"]:
|
||||||
chat.update({
|
chat.update(
|
||||||
"title": chat_new.get("title", chat["title"]),
|
{
|
||||||
"description": chat_new.get("description", chat["description"]),
|
"title": chat_new.get("title", chat["title"]),
|
||||||
"updatedAt": int(datetime.now(tz=timezone.utc).timestamp()),
|
"description": chat_new.get("description", chat["description"]),
|
||||||
"admins": chat_new.get("admins", chat.get("admins") or []),
|
"updatedAt": int(datetime.now(tz=timezone.utc).timestamp()),
|
||||||
"users": chat_new.get("users", chat["users"])
|
"admins": chat_new.get("admins", chat.get("admins") or []),
|
||||||
})
|
"users": chat_new.get("users", chat["users"]),
|
||||||
|
}
|
||||||
|
)
|
||||||
await redis.execute("SET", f"chats/{chat.id}", json.dumps(chat))
|
await redis.execute("SET", f"chats/{chat.id}", json.dumps(chat))
|
||||||
await redis.execute("COMMIT")
|
await redis.execute("COMMIT")
|
||||||
|
|
||||||
return {
|
return {"error": None, "chat": chat}
|
||||||
"error": None,
|
|
||||||
"chat": chat
|
|
||||||
}
|
|
||||||
|
|
||||||
|
|
||||||
@mutation.field("createChat")
|
@mutation.field("createChat")
|
||||||
|
@ -52,7 +49,7 @@ async def update_chat(_, info, chat_new: Chat):
|
||||||
async def create_chat(_, info, title="", members=[]):
|
async def create_chat(_, info, title="", members=[]):
|
||||||
auth: AuthCredentials = info.context["request"].auth
|
auth: AuthCredentials = info.context["request"].auth
|
||||||
chat = {}
|
chat = {}
|
||||||
print('create_chat members: %r' % members)
|
print("create_chat members: %r" % members)
|
||||||
if auth.user_id not in members:
|
if auth.user_id not in members:
|
||||||
members.append(int(auth.user_id))
|
members.append(int(auth.user_id))
|
||||||
|
|
||||||
|
@ -74,15 +71,12 @@ async def create_chat(_, info, title="", members=[]):
|
||||||
chat = await redis.execute("GET", f"chats/{c.decode('utf-8')}")
|
chat = await redis.execute("GET", f"chats/{c.decode('utf-8')}")
|
||||||
if chat:
|
if chat:
|
||||||
chat = json.loads(chat)
|
chat = json.loads(chat)
|
||||||
if chat['title'] == "":
|
if chat["title"] == "":
|
||||||
print('[inbox] createChat found old chat')
|
print("[inbox] createChat found old chat")
|
||||||
print(chat)
|
print(chat)
|
||||||
break
|
break
|
||||||
if chat:
|
if chat:
|
||||||
return {
|
return {"chat": chat, "error": "existed"}
|
||||||
"chat": chat,
|
|
||||||
"error": "existed"
|
|
||||||
}
|
|
||||||
|
|
||||||
chat_id = str(uuid.uuid4())
|
chat_id = str(uuid.uuid4())
|
||||||
chat = {
|
chat = {
|
||||||
|
@ -92,7 +86,7 @@ async def create_chat(_, info, title="", members=[]):
|
||||||
"createdBy": auth.user_id,
|
"createdBy": auth.user_id,
|
||||||
"createdAt": int(datetime.now(tz=timezone.utc).timestamp()),
|
"createdAt": int(datetime.now(tz=timezone.utc).timestamp()),
|
||||||
"updatedAt": int(datetime.now(tz=timezone.utc).timestamp()),
|
"updatedAt": int(datetime.now(tz=timezone.utc).timestamp()),
|
||||||
"admins": members if (len(members) == 2 and title == "") else []
|
"admins": members if (len(members) == 2 and title == "") else [],
|
||||||
}
|
}
|
||||||
|
|
||||||
for m in members:
|
for m in members:
|
||||||
|
@ -100,10 +94,7 @@ async def create_chat(_, info, title="", members=[]):
|
||||||
await redis.execute("SET", f"chats/{chat_id}", json.dumps(chat))
|
await redis.execute("SET", f"chats/{chat_id}", json.dumps(chat))
|
||||||
await redis.execute("SET", f"chats/{chat_id}/next_message_id", str(0))
|
await redis.execute("SET", f"chats/{chat_id}/next_message_id", str(0))
|
||||||
await redis.execute("COMMIT")
|
await redis.execute("COMMIT")
|
||||||
return {
|
return {"error": None, "chat": chat}
|
||||||
"error": None,
|
|
||||||
"chat": chat
|
|
||||||
}
|
|
||||||
|
|
||||||
|
|
||||||
@mutation.field("deleteChat")
|
@mutation.field("deleteChat")
|
||||||
|
@ -114,11 +105,9 @@ async def delete_chat(_, info, chat_id: str):
|
||||||
chat = await redis.execute("GET", f"/chats/{chat_id}")
|
chat = await redis.execute("GET", f"/chats/{chat_id}")
|
||||||
if chat:
|
if chat:
|
||||||
chat = dict(json.loads(chat))
|
chat = dict(json.loads(chat))
|
||||||
if auth.user_id in chat['admins']:
|
if auth.user_id in chat["admins"]:
|
||||||
await redis.execute("DEL", f"chats/{chat_id}")
|
await redis.execute("DEL", f"chats/{chat_id}")
|
||||||
await redis.execute("SREM", "chats_by_user/" + str(auth.user_id), chat_id)
|
await redis.execute("SREM", "chats_by_user/" + str(auth.user_id), chat_id)
|
||||||
await redis.execute("COMMIT")
|
await redis.execute("COMMIT")
|
||||||
else:
|
else:
|
||||||
return {
|
return {"error": "chat not exist"}
|
||||||
"error": "chat not exist"
|
|
||||||
}
|
|
||||||
|
|
|
@ -1,28 +1,27 @@
|
||||||
import json
|
import json
|
||||||
# from datetime import datetime, timedelta, timezone
|
|
||||||
|
|
||||||
from auth.authenticate import login_required
|
from auth.authenticate import login_required
|
||||||
from auth.credentials import AuthCredentials
|
from auth.credentials import AuthCredentials
|
||||||
from base.redis import redis
|
|
||||||
from base.orm import local_session
|
from base.orm import local_session
|
||||||
|
from base.redis import redis
|
||||||
from base.resolvers import query
|
from base.resolvers import query
|
||||||
from orm.user import User
|
from orm.user import User
|
||||||
from resolvers.zine.profile import followed_authors
|
from resolvers.zine.profile import followed_authors
|
||||||
|
|
||||||
from .unread import get_unread_counter
|
from .unread import get_unread_counter
|
||||||
|
|
||||||
|
# from datetime import datetime, timedelta, timezone
|
||||||
|
|
||||||
|
|
||||||
async def load_messages(chat_id: str, limit: int = 5, offset: int = 0, ids=[]):
|
async def load_messages(chat_id: str, limit: int = 5, offset: int = 0, ids=[]):
|
||||||
''' load :limit messages for :chat_id with :offset '''
|
"""load :limit messages for :chat_id with :offset"""
|
||||||
messages = []
|
messages = []
|
||||||
message_ids = []
|
message_ids = []
|
||||||
if ids:
|
if ids:
|
||||||
message_ids += ids
|
message_ids += ids
|
||||||
try:
|
try:
|
||||||
if limit:
|
if limit:
|
||||||
mids = await redis.lrange(f"chats/{chat_id}/message_ids",
|
mids = await redis.lrange(f"chats/{chat_id}/message_ids", offset, offset + limit)
|
||||||
offset,
|
|
||||||
offset + limit
|
|
||||||
)
|
|
||||||
mids = [mid.decode("utf-8") for mid in mids]
|
mids = [mid.decode("utf-8") for mid in mids]
|
||||||
message_ids += mids
|
message_ids += mids
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
|
@ -30,10 +29,10 @@ async def load_messages(chat_id: str, limit: int = 5, offset: int = 0, ids=[]):
|
||||||
if message_ids:
|
if message_ids:
|
||||||
message_keys = [f"chats/{chat_id}/messages/{mid}" for mid in message_ids]
|
message_keys = [f"chats/{chat_id}/messages/{mid}" for mid in message_ids]
|
||||||
messages = await redis.mget(*message_keys)
|
messages = await redis.mget(*message_keys)
|
||||||
messages = [json.loads(msg.decode('utf-8')) for msg in messages]
|
messages = [json.loads(msg.decode("utf-8")) for msg in messages]
|
||||||
replies = []
|
replies = []
|
||||||
for m in messages:
|
for m in messages:
|
||||||
rt = m.get('replyTo')
|
rt = m.get("replyTo")
|
||||||
if rt:
|
if rt:
|
||||||
rt = int(rt)
|
rt = int(rt)
|
||||||
if rt not in message_ids:
|
if rt not in message_ids:
|
||||||
|
@ -46,14 +45,14 @@ async def load_messages(chat_id: str, limit: int = 5, offset: int = 0, ids=[]):
|
||||||
@query.field("loadChats")
|
@query.field("loadChats")
|
||||||
@login_required
|
@login_required
|
||||||
async def load_chats(_, info, limit: int = 50, offset: int = 0):
|
async def load_chats(_, info, limit: int = 50, offset: int = 0):
|
||||||
""" load :limit chats of current user with :offset """
|
"""load :limit chats of current user with :offset"""
|
||||||
auth: AuthCredentials = info.context["request"].auth
|
auth: AuthCredentials = info.context["request"].auth
|
||||||
|
|
||||||
cids = await redis.execute("SMEMBERS", "chats_by_user/" + str(auth.user_id))
|
cids = await redis.execute("SMEMBERS", "chats_by_user/" + str(auth.user_id))
|
||||||
if cids:
|
if cids:
|
||||||
cids = list(cids)[offset:offset + limit]
|
cids = list(cids)[offset : offset + limit]
|
||||||
if not cids:
|
if not cids:
|
||||||
print('[inbox.load] no chats were found')
|
print("[inbox.load] no chats were found")
|
||||||
cids = []
|
cids = []
|
||||||
onliners = await redis.execute("SMEMBERS", "users-online")
|
onliners = await redis.execute("SMEMBERS", "users-online")
|
||||||
if not onliners:
|
if not onliners:
|
||||||
|
@ -64,62 +63,50 @@ async def load_chats(_, info, limit: int = 50, offset: int = 0):
|
||||||
c = await redis.execute("GET", "chats/" + cid)
|
c = await redis.execute("GET", "chats/" + cid)
|
||||||
if c:
|
if c:
|
||||||
c = dict(json.loads(c))
|
c = dict(json.loads(c))
|
||||||
c['messages'] = await load_messages(cid, 5, 0)
|
c["messages"] = await load_messages(cid, 5, 0)
|
||||||
c['unread'] = await get_unread_counter(cid, auth.user_id)
|
c["unread"] = await get_unread_counter(cid, auth.user_id)
|
||||||
with local_session() as session:
|
with local_session() as session:
|
||||||
c['members'] = []
|
c["members"] = []
|
||||||
for uid in c["users"]:
|
for uid in c["users"]:
|
||||||
a = session.query(User).where(User.id == uid).first()
|
a = session.query(User).where(User.id == uid).first()
|
||||||
if a:
|
if a:
|
||||||
c['members'].append({
|
c["members"].append(
|
||||||
"id": a.id,
|
{
|
||||||
"slug": a.slug,
|
"id": a.id,
|
||||||
"userpic": a.userpic,
|
"slug": a.slug,
|
||||||
"name": a.name,
|
"userpic": a.userpic,
|
||||||
"lastSeen": a.lastSeen,
|
"name": a.name,
|
||||||
"online": a.id in onliners
|
"lastSeen": a.lastSeen,
|
||||||
})
|
"online": a.id in onliners,
|
||||||
|
}
|
||||||
|
)
|
||||||
chats.append(c)
|
chats.append(c)
|
||||||
return {
|
return {"chats": chats, "error": None}
|
||||||
"chats": chats,
|
|
||||||
"error": None
|
|
||||||
}
|
|
||||||
|
|
||||||
|
|
||||||
@query.field("loadMessagesBy")
|
@query.field("loadMessagesBy")
|
||||||
@login_required
|
@login_required
|
||||||
async def load_messages_by(_, info, by, limit: int = 10, offset: int = 0):
|
async def load_messages_by(_, info, by, limit: int = 10, offset: int = 0):
|
||||||
''' load :limit messages of :chat_id with :offset '''
|
"""load :limit messages of :chat_id with :offset"""
|
||||||
|
|
||||||
auth: AuthCredentials = info.context["request"].auth
|
auth: AuthCredentials = info.context["request"].auth
|
||||||
userchats = await redis.execute("SMEMBERS", "chats_by_user/" + str(auth.user_id))
|
userchats = await redis.execute("SMEMBERS", "chats_by_user/" + str(auth.user_id))
|
||||||
userchats = [c.decode('utf-8') for c in userchats]
|
userchats = [c.decode("utf-8") for c in userchats]
|
||||||
# print('[inbox] userchats: %r' % userchats)
|
# print('[inbox] userchats: %r' % userchats)
|
||||||
if userchats:
|
if userchats:
|
||||||
# print('[inbox] loading messages by...')
|
# print('[inbox] loading messages by...')
|
||||||
messages = []
|
messages = []
|
||||||
by_chat = by.get('chat')
|
by_chat = by.get("chat")
|
||||||
if by_chat in userchats:
|
if by_chat in userchats:
|
||||||
chat = await redis.execute("GET", f"chats/{by_chat}")
|
chat = await redis.execute("GET", f"chats/{by_chat}")
|
||||||
# print(chat)
|
# print(chat)
|
||||||
if not chat:
|
if not chat:
|
||||||
return {
|
return {"messages": [], "error": "chat not exist"}
|
||||||
"messages": [],
|
|
||||||
"error": "chat not exist"
|
|
||||||
}
|
|
||||||
# everyone's messages in filtered chat
|
# everyone's messages in filtered chat
|
||||||
messages = await load_messages(by_chat, limit, offset)
|
messages = await load_messages(by_chat, limit, offset)
|
||||||
return {
|
return {"messages": sorted(list(messages), key=lambda m: m["createdAt"]), "error": None}
|
||||||
"messages": sorted(
|
|
||||||
list(messages),
|
|
||||||
key=lambda m: m['createdAt']
|
|
||||||
),
|
|
||||||
"error": None
|
|
||||||
}
|
|
||||||
else:
|
else:
|
||||||
return {
|
return {"error": "Cannot access messages of this chat"}
|
||||||
"error": "Cannot access messages of this chat"
|
|
||||||
}
|
|
||||||
|
|
||||||
|
|
||||||
@query.field("loadRecipients")
|
@query.field("loadRecipients")
|
||||||
|
@ -138,15 +125,14 @@ async def load_recipients(_, info, limit=50, offset=0):
|
||||||
chat_users += session.query(User).where(User.emailConfirmed).limit(limit).offset(offset)
|
chat_users += session.query(User).where(User.emailConfirmed).limit(limit).offset(offset)
|
||||||
members = []
|
members = []
|
||||||
for a in chat_users:
|
for a in chat_users:
|
||||||
members.append({
|
members.append(
|
||||||
"id": a.id,
|
{
|
||||||
"slug": a.slug,
|
"id": a.id,
|
||||||
"userpic": a.userpic,
|
"slug": a.slug,
|
||||||
"name": a.name,
|
"userpic": a.userpic,
|
||||||
"lastSeen": a.lastSeen,
|
"name": a.name,
|
||||||
"online": a.id in onliners
|
"lastSeen": a.lastSeen,
|
||||||
})
|
"online": a.id in onliners,
|
||||||
return {
|
}
|
||||||
"members": members,
|
)
|
||||||
"error": None
|
return {"members": members, "error": None}
|
||||||
}
|
|
||||||
|
|
|
@ -1,62 +1,54 @@
|
||||||
import asyncio
|
|
||||||
import json
|
import json
|
||||||
from typing import Any
|
|
||||||
from datetime import datetime, timezone
|
from datetime import datetime, timezone
|
||||||
from graphql.type import GraphQLResolveInfo
|
|
||||||
from auth.authenticate import login_required
|
from auth.authenticate import login_required
|
||||||
from auth.credentials import AuthCredentials
|
from auth.credentials import AuthCredentials
|
||||||
from base.redis import redis
|
from base.redis import redis
|
||||||
from base.resolvers import mutation
|
from base.resolvers import mutation
|
||||||
from services.following import FollowingManager, FollowingResult, Following
|
from services.following import FollowingManager, FollowingResult
|
||||||
from validations.inbox import Message
|
|
||||||
|
|
||||||
|
|
||||||
@mutation.field("createMessage")
|
@mutation.field("createMessage")
|
||||||
@login_required
|
@login_required
|
||||||
async def create_message(_, info, chat: str, body: str, replyTo=None):
|
async def create_message(_, info, chat: str, body: str, replyTo=None):
|
||||||
""" create message with :body for :chat_id replying to :replyTo optionally """
|
"""create message with :body for :chat_id replying to :replyTo optionally"""
|
||||||
auth: AuthCredentials = info.context["request"].auth
|
auth: AuthCredentials = info.context["request"].auth
|
||||||
|
|
||||||
chat = await redis.execute("GET", f"chats/{chat}")
|
chat = await redis.execute("GET", f"chats/{chat}")
|
||||||
if not chat:
|
if not chat:
|
||||||
return {
|
return {"error": "chat is not exist"}
|
||||||
"error": "chat is not exist"
|
|
||||||
}
|
|
||||||
else:
|
else:
|
||||||
chat = dict(json.loads(chat))
|
chat_dict = dict(json.loads(chat))
|
||||||
message_id = await redis.execute("GET", f"chats/{chat['id']}/next_message_id")
|
message_id = await redis.execute("GET", f"chats/{chat_dict['id']}/next_message_id")
|
||||||
message_id = int(message_id)
|
message_id = int(message_id)
|
||||||
new_message = {
|
new_message = {
|
||||||
"chatId": chat['id'],
|
"chatId": chat_dict["id"],
|
||||||
"id": message_id,
|
"id": message_id,
|
||||||
"author": auth.user_id,
|
"author": auth.user_id,
|
||||||
"body": body,
|
"body": body,
|
||||||
"createdAt": int(datetime.now(tz=timezone.utc).timestamp())
|
"createdAt": int(datetime.now(tz=timezone.utc).timestamp()),
|
||||||
}
|
}
|
||||||
if replyTo:
|
if replyTo:
|
||||||
new_message['replyTo'] = replyTo
|
new_message["replyTo"] = replyTo
|
||||||
chat['updatedAt'] = new_message['createdAt']
|
chat_dict["updatedAt"] = new_message["createdAt"]
|
||||||
await redis.execute("SET", f"chats/{chat['id']}", json.dumps(chat))
|
await redis.execute("SET", f"chats/{chat_dict['id']}", json.dumps(chat))
|
||||||
print(f"[inbox] creating message {new_message}")
|
print(f"[inbox] creating message {new_message}")
|
||||||
await redis.execute(
|
await redis.execute(
|
||||||
"SET", f"chats/{chat['id']}/messages/{message_id}", json.dumps(new_message)
|
"SET", f"chats/{chat_dict['id']}/messages/{message_id}", json.dumps(new_message)
|
||||||
)
|
)
|
||||||
await redis.execute("LPUSH", f"chats/{chat['id']}/message_ids", str(message_id))
|
await redis.execute("LPUSH", f"chats/{chat_dict['id']}/message_ids", str(message_id))
|
||||||
await redis.execute("SET", f"chats/{chat['id']}/next_message_id", str(message_id + 1))
|
await redis.execute("SET", f"chats/{chat_dict['id']}/next_message_id", str(message_id + 1))
|
||||||
|
|
||||||
users = chat["users"]
|
users = chat_dict["users"]
|
||||||
for user_slug in users:
|
for user_slug in users:
|
||||||
await redis.execute(
|
await redis.execute(
|
||||||
"LPUSH", f"chats/{chat['id']}/unread/{user_slug}", str(message_id)
|
"LPUSH", f"chats/{chat_dict['id']}/unread/{user_slug}", str(message_id)
|
||||||
)
|
)
|
||||||
|
|
||||||
result = FollowingResult("NEW", 'chat', new_message)
|
result = FollowingResult("NEW", "chat", new_message)
|
||||||
await FollowingManager.push('chat', result)
|
await FollowingManager.push("chat", result)
|
||||||
|
|
||||||
return {
|
return {"message": new_message, "error": None}
|
||||||
"message": new_message,
|
|
||||||
"error": None
|
|
||||||
}
|
|
||||||
|
|
||||||
|
|
||||||
@mutation.field("updateMessage")
|
@mutation.field("updateMessage")
|
||||||
|
@ -81,13 +73,10 @@ async def update_message(_, info, chat_id: str, message_id: int, body: str):
|
||||||
|
|
||||||
await redis.execute("SET", f"chats/{chat_id}/messages/{message_id}", json.dumps(message))
|
await redis.execute("SET", f"chats/{chat_id}/messages/{message_id}", json.dumps(message))
|
||||||
|
|
||||||
result = FollowingResult("UPDATED", 'chat', message)
|
result = FollowingResult("UPDATED", "chat", message)
|
||||||
await FollowingManager.push('chat', result)
|
await FollowingManager.push("chat", result)
|
||||||
|
|
||||||
return {
|
return {"message": message, "error": None}
|
||||||
"message": message,
|
|
||||||
"error": None
|
|
||||||
}
|
|
||||||
|
|
||||||
|
|
||||||
@mutation.field("deleteMessage")
|
@mutation.field("deleteMessage")
|
||||||
|
@ -114,7 +103,7 @@ async def delete_message(_, info, chat_id: str, message_id: int):
|
||||||
for user_id in users:
|
for user_id in users:
|
||||||
await redis.execute("LREM", f"chats/{chat_id}/unread/{user_id}", 0, str(message_id))
|
await redis.execute("LREM", f"chats/{chat_id}/unread/{user_id}", 0, str(message_id))
|
||||||
|
|
||||||
result = FollowingResult("DELETED", 'chat', message)
|
result = FollowingResult("DELETED", "chat", message)
|
||||||
await FollowingManager.push(result)
|
await FollowingManager.push(result)
|
||||||
|
|
||||||
return {}
|
return {}
|
||||||
|
@ -137,6 +126,4 @@ async def mark_as_read(_, info, chat_id: str, messages: [int]):
|
||||||
for message_id in messages:
|
for message_id in messages:
|
||||||
await redis.execute("LREM", f"chats/{chat_id}/unread/{auth.user_id}", 0, str(message_id))
|
await redis.execute("LREM", f"chats/{chat_id}/unread/{auth.user_id}", 0, str(message_id))
|
||||||
|
|
||||||
return {
|
return {"error": None}
|
||||||
"error": None
|
|
||||||
}
|
|
||||||
|
|
|
@ -1,10 +1,11 @@
|
||||||
import json
|
import json
|
||||||
from datetime import datetime, timezone, timedelta
|
from datetime import datetime, timedelta, timezone
|
||||||
|
|
||||||
from auth.authenticate import login_required
|
from auth.authenticate import login_required
|
||||||
from auth.credentials import AuthCredentials
|
from auth.credentials import AuthCredentials
|
||||||
|
from base.orm import local_session
|
||||||
from base.redis import redis
|
from base.redis import redis
|
||||||
from base.resolvers import query
|
from base.resolvers import query
|
||||||
from base.orm import local_session
|
|
||||||
from orm.user import AuthorFollower, User
|
from orm.user import AuthorFollower, User
|
||||||
from resolvers.inbox.load import load_messages
|
from resolvers.inbox.load import load_messages
|
||||||
|
|
||||||
|
@ -17,7 +18,7 @@ async def search_recipients(_, info, query: str, limit: int = 50, offset: int =
|
||||||
auth: AuthCredentials = info.context["request"].auth
|
auth: AuthCredentials = info.context["request"].auth
|
||||||
talk_before = await redis.execute("GET", f"/chats_by_user/{auth.user_id}")
|
talk_before = await redis.execute("GET", f"/chats_by_user/{auth.user_id}")
|
||||||
if talk_before:
|
if talk_before:
|
||||||
talk_before = list(json.loads(talk_before))[offset:offset + limit]
|
talk_before = list(json.loads(talk_before))[offset : offset + limit]
|
||||||
for chat_id in talk_before:
|
for chat_id in talk_before:
|
||||||
members = await redis.execute("GET", f"/chats/{chat_id}/users")
|
members = await redis.execute("GET", f"/chats/{chat_id}/users")
|
||||||
if members:
|
if members:
|
||||||
|
@ -31,23 +32,24 @@ async def search_recipients(_, info, query: str, limit: int = 50, offset: int =
|
||||||
|
|
||||||
with local_session() as session:
|
with local_session() as session:
|
||||||
# followings
|
# followings
|
||||||
result += session.query(AuthorFollower.author).join(
|
result += (
|
||||||
User, User.id == AuthorFollower.follower
|
session.query(AuthorFollower.author)
|
||||||
).where(
|
.join(User, User.id == AuthorFollower.follower)
|
||||||
User.slug.startswith(query)
|
.where(User.slug.startswith(query))
|
||||||
).offset(offset + len(result)).limit(more_amount)
|
.offset(offset + len(result))
|
||||||
|
.limit(more_amount)
|
||||||
|
)
|
||||||
|
|
||||||
more_amount = limit
|
more_amount = limit
|
||||||
# followers
|
# followers
|
||||||
result += session.query(AuthorFollower.follower).join(
|
result += (
|
||||||
User, User.id == AuthorFollower.author
|
session.query(AuthorFollower.follower)
|
||||||
).where(
|
.join(User, User.id == AuthorFollower.author)
|
||||||
User.slug.startswith(query)
|
.where(User.slug.startswith(query))
|
||||||
).offset(offset + len(result)).limit(offset + len(result) + limit)
|
.offset(offset + len(result))
|
||||||
return {
|
.limit(offset + len(result) + limit)
|
||||||
"members": list(result),
|
)
|
||||||
"error": None
|
return {"members": list(result), "error": None}
|
||||||
}
|
|
||||||
|
|
||||||
|
|
||||||
@query.field("searchMessages")
|
@query.field("searchMessages")
|
||||||
|
@ -57,22 +59,22 @@ async def search_user_chats(by, messages, user_id: int, limit, offset):
|
||||||
cids.union(set(await redis.execute("SMEMBERS", "chats_by_user/" + str(user_id))))
|
cids.union(set(await redis.execute("SMEMBERS", "chats_by_user/" + str(user_id))))
|
||||||
messages = []
|
messages = []
|
||||||
|
|
||||||
by_author = by.get('author')
|
by_author = by.get("author")
|
||||||
if by_author:
|
if by_author:
|
||||||
# all author's messages
|
# all author's messages
|
||||||
cids.union(set(await redis.execute("SMEMBERS", f"chats_by_user/{by_author}")))
|
cids.union(set(await redis.execute("SMEMBERS", f"chats_by_user/{by_author}")))
|
||||||
# author's messages in filtered chat
|
# author's messages in filtered chat
|
||||||
messages.union(set(filter(lambda m: m["author"] == by_author, list(messages))))
|
messages.union(set(filter(lambda m: m["author"] == by_author, list(messages))))
|
||||||
for c in cids:
|
for c in cids:
|
||||||
c = c.decode('utf-8')
|
c = c.decode("utf-8")
|
||||||
messages = await load_messages(c, limit, offset)
|
messages = await load_messages(c, limit, offset)
|
||||||
|
|
||||||
body_like = by.get('body')
|
body_like = by.get("body")
|
||||||
if body_like:
|
if body_like:
|
||||||
# search in all messages in all user's chats
|
# search in all messages in all user's chats
|
||||||
for c in cids:
|
for c in cids:
|
||||||
# FIXME: use redis scan here
|
# FIXME: use redis scan here
|
||||||
c = c.decode('utf-8')
|
c = c.decode("utf-8")
|
||||||
mmm = await load_messages(c, limit, offset)
|
mmm = await load_messages(c, limit, offset)
|
||||||
for m in mmm:
|
for m in mmm:
|
||||||
if body_like in m["body"]:
|
if body_like in m["body"]:
|
||||||
|
@ -83,13 +85,12 @@ async def search_user_chats(by, messages, user_id: int, limit, offset):
|
||||||
|
|
||||||
days = by.get("days")
|
days = by.get("days")
|
||||||
if days:
|
if days:
|
||||||
messages.extend(filter(
|
messages.extend(
|
||||||
list(messages),
|
filter(
|
||||||
key=lambda m: (
|
list(messages),
|
||||||
datetime.now(tz=timezone.utc) - int(m["createdAt"]) < timedelta(days=by["days"])
|
key=lambda m: (
|
||||||
|
datetime.now(tz=timezone.utc) - int(m["createdAt"]) < timedelta(days=by["days"])
|
||||||
|
),
|
||||||
)
|
)
|
||||||
))
|
)
|
||||||
return {
|
return {"messages": messages, "error": None}
|
||||||
"messages": messages,
|
|
||||||
"error": None
|
|
||||||
}
|
|
||||||
|
|
|
@ -1,9 +1,9 @@
|
||||||
from sqlalchemy import select, desc, and_, update
|
from sqlalchemy import and_, desc, select, update
|
||||||
|
|
||||||
from auth.credentials import AuthCredentials
|
|
||||||
from base.resolvers import query, mutation
|
|
||||||
from auth.authenticate import login_required
|
from auth.authenticate import login_required
|
||||||
|
from auth.credentials import AuthCredentials
|
||||||
from base.orm import local_session
|
from base.orm import local_session
|
||||||
|
from base.resolvers import mutation, query
|
||||||
from orm import Notification
|
from orm import Notification
|
||||||
|
|
||||||
|
|
||||||
|
@ -16,25 +16,26 @@ async def load_notifications(_, info, params=None):
|
||||||
auth: AuthCredentials = info.context["request"].auth
|
auth: AuthCredentials = info.context["request"].auth
|
||||||
user_id = auth.user_id
|
user_id = auth.user_id
|
||||||
|
|
||||||
limit = params.get('limit', 50)
|
limit = params.get("limit", 50)
|
||||||
offset = params.get('offset', 0)
|
offset = params.get("offset", 0)
|
||||||
|
|
||||||
q = select(Notification).where(
|
q = (
|
||||||
Notification.user == user_id
|
select(Notification)
|
||||||
).order_by(desc(Notification.createdAt)).limit(limit).offset(offset)
|
.where(Notification.user == user_id)
|
||||||
|
.order_by(desc(Notification.createdAt))
|
||||||
|
.limit(limit)
|
||||||
|
.offset(offset)
|
||||||
|
)
|
||||||
|
|
||||||
notifications = []
|
notifications = []
|
||||||
with local_session() as session:
|
with local_session() as session:
|
||||||
total_count = session.query(Notification).where(
|
total_count = session.query(Notification).where(Notification.user == user_id).count()
|
||||||
Notification.user == user_id
|
|
||||||
).count()
|
|
||||||
|
|
||||||
total_unread_count = session.query(Notification).where(
|
total_unread_count = (
|
||||||
and_(
|
session.query(Notification)
|
||||||
Notification.user == user_id,
|
.where(and_(Notification.user == user_id, Notification.seen == False)) # noqa: E712
|
||||||
Notification.seen == False
|
.count()
|
||||||
)
|
)
|
||||||
).count()
|
|
||||||
|
|
||||||
for [notification] in session.execute(q):
|
for [notification] in session.execute(q):
|
||||||
notification.type = notification.type.name
|
notification.type = notification.type.name
|
||||||
|
@ -43,7 +44,7 @@ async def load_notifications(_, info, params=None):
|
||||||
return {
|
return {
|
||||||
"notifications": notifications,
|
"notifications": notifications,
|
||||||
"totalCount": total_count,
|
"totalCount": total_count,
|
||||||
"totalUnreadCount": total_unread_count
|
"totalUnreadCount": total_unread_count,
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
|
@ -54,9 +55,11 @@ async def mark_notification_as_read(_, info, notification_id: int):
|
||||||
user_id = auth.user_id
|
user_id = auth.user_id
|
||||||
|
|
||||||
with local_session() as session:
|
with local_session() as session:
|
||||||
notification = session.query(Notification).where(
|
notification = (
|
||||||
and_(Notification.id == notification_id, Notification.user == user_id)
|
session.query(Notification)
|
||||||
).one()
|
.where(and_(Notification.id == notification_id, Notification.user == user_id))
|
||||||
|
.one()
|
||||||
|
)
|
||||||
notification.seen = True
|
notification.seen = True
|
||||||
session.commit()
|
session.commit()
|
||||||
|
|
||||||
|
@ -69,12 +72,11 @@ async def mark_all_notifications_as_read(_, info):
|
||||||
auth: AuthCredentials = info.context["request"].auth
|
auth: AuthCredentials = info.context["request"].auth
|
||||||
user_id = auth.user_id
|
user_id = auth.user_id
|
||||||
|
|
||||||
statement = update(Notification).where(
|
statement = (
|
||||||
and_(
|
update(Notification)
|
||||||
Notification.user == user_id,
|
.where(and_(Notification.user == user_id, Notification.seen == False)) # noqa: E712
|
||||||
Notification.seen == False
|
.values(seen=True)
|
||||||
)
|
)
|
||||||
).values(seen=True)
|
|
||||||
|
|
||||||
with local_session() as session:
|
with local_session() as session:
|
||||||
try:
|
try:
|
||||||
|
|
|
@ -2,33 +2,36 @@ import os
|
||||||
import shutil
|
import shutil
|
||||||
import tempfile
|
import tempfile
|
||||||
import uuid
|
import uuid
|
||||||
|
|
||||||
import boto3
|
import boto3
|
||||||
from botocore.exceptions import BotoCoreError, ClientError
|
from botocore.exceptions import BotoCoreError, ClientError
|
||||||
from starlette.responses import JSONResponse
|
from starlette.responses import JSONResponse
|
||||||
|
|
||||||
STORJ_ACCESS_KEY = os.environ.get('STORJ_ACCESS_KEY')
|
STORJ_ACCESS_KEY = os.environ.get("STORJ_ACCESS_KEY")
|
||||||
STORJ_SECRET_KEY = os.environ.get('STORJ_SECRET_KEY')
|
STORJ_SECRET_KEY = os.environ.get("STORJ_SECRET_KEY")
|
||||||
STORJ_END_POINT = os.environ.get('STORJ_END_POINT')
|
STORJ_END_POINT = os.environ.get("STORJ_END_POINT")
|
||||||
STORJ_BUCKET_NAME = os.environ.get('STORJ_BUCKET_NAME')
|
STORJ_BUCKET_NAME = os.environ.get("STORJ_BUCKET_NAME")
|
||||||
CDN_DOMAIN = os.environ.get('CDN_DOMAIN')
|
CDN_DOMAIN = os.environ.get("CDN_DOMAIN")
|
||||||
|
|
||||||
|
|
||||||
async def upload_handler(request):
|
async def upload_handler(request):
|
||||||
form = await request.form()
|
form = await request.form()
|
||||||
file = form.get('file')
|
file = form.get("file")
|
||||||
|
|
||||||
if file is None:
|
if file is None:
|
||||||
return JSONResponse({'error': 'No file uploaded'}, status_code=400)
|
return JSONResponse({"error": "No file uploaded"}, status_code=400)
|
||||||
|
|
||||||
file_name, file_extension = os.path.splitext(file.filename)
|
file_name, file_extension = os.path.splitext(file.filename)
|
||||||
|
|
||||||
key = 'files/' + str(uuid.uuid4()) + file_extension
|
key = "files/" + str(uuid.uuid4()) + file_extension
|
||||||
|
|
||||||
# Create an S3 client with Storj configuration
|
# Create an S3 client with Storj configuration
|
||||||
s3 = boto3.client('s3',
|
s3 = boto3.client(
|
||||||
aws_access_key_id=STORJ_ACCESS_KEY,
|
"s3",
|
||||||
aws_secret_access_key=STORJ_SECRET_KEY,
|
aws_access_key_id=STORJ_ACCESS_KEY,
|
||||||
endpoint_url=STORJ_END_POINT)
|
aws_secret_access_key=STORJ_SECRET_KEY,
|
||||||
|
endpoint_url=STORJ_END_POINT,
|
||||||
|
)
|
||||||
|
|
||||||
try:
|
try:
|
||||||
# Save the uploaded file to a temporary file
|
# Save the uploaded file to a temporary file
|
||||||
|
@ -39,18 +42,13 @@ async def upload_handler(request):
|
||||||
Filename=tmp_file.name,
|
Filename=tmp_file.name,
|
||||||
Bucket=STORJ_BUCKET_NAME,
|
Bucket=STORJ_BUCKET_NAME,
|
||||||
Key=key,
|
Key=key,
|
||||||
ExtraArgs={
|
ExtraArgs={"ContentType": file.content_type},
|
||||||
"ContentType": file.content_type
|
|
||||||
}
|
|
||||||
)
|
)
|
||||||
|
|
||||||
url = 'https://' + CDN_DOMAIN + '/' + key
|
url = "https://" + CDN_DOMAIN + "/" + key
|
||||||
|
|
||||||
return JSONResponse({'url': url, 'originalFilename': file.filename})
|
return JSONResponse({"url": url, "originalFilename": file.filename})
|
||||||
|
|
||||||
except (BotoCoreError, ClientError) as e:
|
except (BotoCoreError, ClientError) as e:
|
||||||
print(e)
|
print(e)
|
||||||
return JSONResponse({'error': 'Failed to upload file'}, status_code=500)
|
return JSONResponse({"error": "Failed to upload file"}, status_code=500)
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
|
|
@ -1,41 +1,36 @@
|
||||||
import asyncio
|
|
||||||
from base.orm import local_session
|
|
||||||
from base.resolvers import mutation
|
|
||||||
from auth.authenticate import login_required
|
from auth.authenticate import login_required
|
||||||
from auth.credentials import AuthCredentials
|
from auth.credentials import AuthCredentials
|
||||||
|
from base.resolvers import mutation
|
||||||
|
|
||||||
# from resolvers.community import community_follow, community_unfollow
|
# from resolvers.community import community_follow, community_unfollow
|
||||||
from orm.user import AuthorFollower
|
|
||||||
from orm.topic import TopicFollower
|
|
||||||
from orm.shout import ShoutReactionsFollower
|
|
||||||
from resolvers.zine.profile import author_follow, author_unfollow
|
from resolvers.zine.profile import author_follow, author_unfollow
|
||||||
from resolvers.zine.reactions import reactions_follow, reactions_unfollow
|
from resolvers.zine.reactions import reactions_follow, reactions_unfollow
|
||||||
from resolvers.zine.topics import topic_follow, topic_unfollow
|
from resolvers.zine.topics import topic_follow, topic_unfollow
|
||||||
from services.following import Following, FollowingManager, FollowingResult
|
from services.following import FollowingManager, FollowingResult
|
||||||
from graphql.type import GraphQLResolveInfo
|
|
||||||
|
|
||||||
|
|
||||||
@mutation.field("follow")
|
@mutation.field("follow")
|
||||||
@login_required
|
@login_required
|
||||||
async def follow(_, info, what, slug):
|
async def follow(_, info, what, slug): # noqa: C901
|
||||||
auth: AuthCredentials = info.context["request"].auth
|
auth: AuthCredentials = info.context["request"].auth
|
||||||
|
|
||||||
try:
|
try:
|
||||||
if what == "AUTHOR":
|
if what == "AUTHOR":
|
||||||
if author_follow(auth.user_id, slug):
|
if author_follow(auth.user_id, slug):
|
||||||
result = FollowingResult("NEW", 'author', slug)
|
result = FollowingResult("NEW", "author", slug)
|
||||||
await FollowingManager.push('author', result)
|
await FollowingManager.push("author", result)
|
||||||
elif what == "TOPIC":
|
elif what == "TOPIC":
|
||||||
if topic_follow(auth.user_id, slug):
|
if topic_follow(auth.user_id, slug):
|
||||||
result = FollowingResult("NEW", 'topic', slug)
|
result = FollowingResult("NEW", "topic", slug)
|
||||||
await FollowingManager.push('topic', result)
|
await FollowingManager.push("topic", result)
|
||||||
elif what == "COMMUNITY":
|
elif what == "COMMUNITY":
|
||||||
if False: # TODO: use community_follow(auth.user_id, slug):
|
if False: # TODO: use community_follow(auth.user_id, slug):
|
||||||
result = FollowingResult("NEW", 'community', slug)
|
result = FollowingResult("NEW", "community", slug)
|
||||||
await FollowingManager.push('community', result)
|
await FollowingManager.push("community", result)
|
||||||
elif what == "REACTIONS":
|
elif what == "REACTIONS":
|
||||||
if reactions_follow(auth.user_id, slug):
|
if reactions_follow(auth.user_id, slug):
|
||||||
result = FollowingResult("NEW", 'shout', slug)
|
result = FollowingResult("NEW", "shout", slug)
|
||||||
await FollowingManager.push('shout', result)
|
await FollowingManager.push("shout", result)
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
print(Exception(e))
|
print(Exception(e))
|
||||||
return {"error": str(e)}
|
return {"error": str(e)}
|
||||||
|
@ -45,26 +40,26 @@ async def follow(_, info, what, slug):
|
||||||
|
|
||||||
@mutation.field("unfollow")
|
@mutation.field("unfollow")
|
||||||
@login_required
|
@login_required
|
||||||
async def unfollow(_, info, what, slug):
|
async def unfollow(_, info, what, slug): # noqa: C901
|
||||||
auth: AuthCredentials = info.context["request"].auth
|
auth: AuthCredentials = info.context["request"].auth
|
||||||
|
|
||||||
try:
|
try:
|
||||||
if what == "AUTHOR":
|
if what == "AUTHOR":
|
||||||
if author_unfollow(auth.user_id, slug):
|
if author_unfollow(auth.user_id, slug):
|
||||||
result = FollowingResult("DELETED", 'author', slug)
|
result = FollowingResult("DELETED", "author", slug)
|
||||||
await FollowingManager.push('author', result)
|
await FollowingManager.push("author", result)
|
||||||
elif what == "TOPIC":
|
elif what == "TOPIC":
|
||||||
if topic_unfollow(auth.user_id, slug):
|
if topic_unfollow(auth.user_id, slug):
|
||||||
result = FollowingResult("DELETED", 'topic', slug)
|
result = FollowingResult("DELETED", "topic", slug)
|
||||||
await FollowingManager.push('topic', result)
|
await FollowingManager.push("topic", result)
|
||||||
elif what == "COMMUNITY":
|
elif what == "COMMUNITY":
|
||||||
if False: # TODO: use community_unfollow(auth.user_id, slug):
|
if False: # TODO: use community_unfollow(auth.user_id, slug):
|
||||||
result = FollowingResult("DELETED", 'community', slug)
|
result = FollowingResult("DELETED", "community", slug)
|
||||||
await FollowingManager.push('community', result)
|
await FollowingManager.push("community", result)
|
||||||
elif what == "REACTIONS":
|
elif what == "REACTIONS":
|
||||||
if reactions_unfollow(auth.user_id, slug):
|
if reactions_unfollow(auth.user_id, slug):
|
||||||
result = FollowingResult("DELETED", 'shout', slug)
|
result = FollowingResult("DELETED", "shout", slug)
|
||||||
await FollowingManager.push('shout', result)
|
await FollowingManager.push("shout", result)
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
return {"error": str(e)}
|
return {"error": str(e)}
|
||||||
|
|
||||||
|
|
|
@ -1,11 +1,11 @@
|
||||||
from datetime import datetime, timedelta, timezone
|
from datetime import datetime, timedelta, timezone
|
||||||
|
|
||||||
from sqlalchemy.orm import joinedload, aliased
|
from sqlalchemy.orm import aliased, joinedload
|
||||||
from sqlalchemy.sql.expression import desc, asc, select, func, case, and_, text, nulls_last
|
from sqlalchemy.sql.expression import and_, asc, case, desc, func, nulls_last, select
|
||||||
|
|
||||||
from auth.authenticate import login_required
|
from auth.authenticate import login_required
|
||||||
from auth.credentials import AuthCredentials
|
from auth.credentials import AuthCredentials
|
||||||
from base.exceptions import ObjectNotExist, OperationNotAllowed
|
from base.exceptions import ObjectNotExist
|
||||||
from base.orm import local_session
|
from base.orm import local_session
|
||||||
from base.resolvers import query
|
from base.resolvers import query
|
||||||
from orm import TopicFollower
|
from orm import TopicFollower
|
||||||
|
@ -18,37 +18,37 @@ def add_stat_columns(q):
|
||||||
aliased_reaction = aliased(Reaction)
|
aliased_reaction = aliased(Reaction)
|
||||||
|
|
||||||
q = q.outerjoin(aliased_reaction).add_columns(
|
q = q.outerjoin(aliased_reaction).add_columns(
|
||||||
func.sum(
|
func.sum(aliased_reaction.id).label("reacted_stat"),
|
||||||
aliased_reaction.id
|
func.sum(case((aliased_reaction.kind == ReactionKind.COMMENT, 1), else_=0)).label(
|
||||||
).label('reacted_stat'),
|
"commented_stat"
|
||||||
|
),
|
||||||
func.sum(
|
func.sum(
|
||||||
case(
|
case(
|
||||||
(aliased_reaction.kind == ReactionKind.COMMENT, 1),
|
# do not count comments' reactions
|
||||||
else_=0
|
(aliased_reaction.replyTo.is_not(None), 0),
|
||||||
|
(aliased_reaction.kind == ReactionKind.AGREE, 1),
|
||||||
|
(aliased_reaction.kind == ReactionKind.DISAGREE, -1),
|
||||||
|
(aliased_reaction.kind == ReactionKind.PROOF, 1),
|
||||||
|
(aliased_reaction.kind == ReactionKind.DISPROOF, -1),
|
||||||
|
(aliased_reaction.kind == ReactionKind.ACCEPT, 1),
|
||||||
|
(aliased_reaction.kind == ReactionKind.REJECT, -1),
|
||||||
|
(aliased_reaction.kind == ReactionKind.LIKE, 1),
|
||||||
|
(aliased_reaction.kind == ReactionKind.DISLIKE, -1),
|
||||||
|
else_=0,
|
||||||
)
|
)
|
||||||
).label('commented_stat'),
|
).label("rating_stat"),
|
||||||
func.sum(case(
|
func.max(
|
||||||
# do not count comments' reactions
|
case(
|
||||||
(aliased_reaction.replyTo.is_not(None), 0),
|
(aliased_reaction.kind != ReactionKind.COMMENT, None),
|
||||||
(aliased_reaction.kind == ReactionKind.AGREE, 1),
|
else_=aliased_reaction.createdAt,
|
||||||
(aliased_reaction.kind == ReactionKind.DISAGREE, -1),
|
)
|
||||||
(aliased_reaction.kind == ReactionKind.PROOF, 1),
|
).label("last_comment"),
|
||||||
(aliased_reaction.kind == ReactionKind.DISPROOF, -1),
|
)
|
||||||
(aliased_reaction.kind == ReactionKind.ACCEPT, 1),
|
|
||||||
(aliased_reaction.kind == ReactionKind.REJECT, -1),
|
|
||||||
(aliased_reaction.kind == ReactionKind.LIKE, 1),
|
|
||||||
(aliased_reaction.kind == ReactionKind.DISLIKE, -1),
|
|
||||||
else_=0)
|
|
||||||
).label('rating_stat'),
|
|
||||||
func.max(case(
|
|
||||||
(aliased_reaction.kind != ReactionKind.COMMENT, None),
|
|
||||||
else_=aliased_reaction.createdAt
|
|
||||||
)).label('last_comment'))
|
|
||||||
|
|
||||||
return q
|
return q
|
||||||
|
|
||||||
|
|
||||||
def apply_filters(q, filters, user_id=None):
|
def apply_filters(q, filters, user_id=None): # noqa: C901
|
||||||
if filters.get("reacted") and user_id:
|
if filters.get("reacted") and user_id:
|
||||||
q.join(Reaction, Reaction.createdBy == user_id)
|
q.join(Reaction, Reaction.createdBy == user_id)
|
||||||
|
|
||||||
|
@ -60,7 +60,7 @@ def apply_filters(q, filters, user_id=None):
|
||||||
|
|
||||||
if filters.get("layout"):
|
if filters.get("layout"):
|
||||||
q = q.filter(Shout.layout == filters.get("layout"))
|
q = q.filter(Shout.layout == filters.get("layout"))
|
||||||
if filters.get('excludeLayout'):
|
if filters.get("excludeLayout"):
|
||||||
q = q.filter(Shout.layout != filters.get("excludeLayout"))
|
q = q.filter(Shout.layout != filters.get("excludeLayout"))
|
||||||
if filters.get("author"):
|
if filters.get("author"):
|
||||||
q = q.filter(Shout.authors.any(slug=filters.get("author")))
|
q = q.filter(Shout.authors.any(slug=filters.get("author")))
|
||||||
|
@ -87,27 +87,23 @@ async def load_shout(_, info, slug=None, shout_id=None):
|
||||||
q = add_stat_columns(q)
|
q = add_stat_columns(q)
|
||||||
|
|
||||||
if slug is not None:
|
if slug is not None:
|
||||||
q = q.filter(
|
q = q.filter(Shout.slug == slug)
|
||||||
Shout.slug == slug
|
|
||||||
)
|
|
||||||
|
|
||||||
if shout_id is not None:
|
if shout_id is not None:
|
||||||
q = q.filter(
|
q = q.filter(Shout.id == shout_id)
|
||||||
Shout.id == shout_id
|
|
||||||
)
|
|
||||||
|
|
||||||
q = q.filter(
|
q = q.filter(Shout.deletedAt.is_(None)).group_by(Shout.id)
|
||||||
Shout.deletedAt.is_(None)
|
|
||||||
).group_by(Shout.id)
|
|
||||||
|
|
||||||
try:
|
try:
|
||||||
[shout, reacted_stat, commented_stat, rating_stat, last_comment] = session.execute(q).first()
|
[shout, reacted_stat, commented_stat, rating_stat, last_comment] = session.execute(
|
||||||
|
q
|
||||||
|
).first()
|
||||||
|
|
||||||
shout.stat = {
|
shout.stat = {
|
||||||
"viewed": shout.views,
|
"viewed": shout.views,
|
||||||
"reacted": reacted_stat,
|
"reacted": reacted_stat,
|
||||||
"commented": commented_stat,
|
"commented": commented_stat,
|
||||||
"rating": rating_stat
|
"rating": rating_stat,
|
||||||
}
|
}
|
||||||
|
|
||||||
for author_caption in session.query(ShoutAuthor).join(Shout).where(Shout.slug == slug):
|
for author_caption in session.query(ShoutAuthor).join(Shout).where(Shout.slug == slug):
|
||||||
|
@ -142,14 +138,13 @@ async def load_shouts_by(_, info, options):
|
||||||
:return: Shout[]
|
:return: Shout[]
|
||||||
"""
|
"""
|
||||||
|
|
||||||
q = select(Shout).options(
|
q = (
|
||||||
joinedload(Shout.authors),
|
select(Shout)
|
||||||
joinedload(Shout.topics),
|
.options(
|
||||||
).where(
|
joinedload(Shout.authors),
|
||||||
and_(
|
joinedload(Shout.topics),
|
||||||
Shout.deletedAt.is_(None),
|
|
||||||
Shout.layout.is_not(None)
|
|
||||||
)
|
)
|
||||||
|
.where(and_(Shout.deletedAt.is_(None), Shout.layout.is_not(None)))
|
||||||
)
|
)
|
||||||
|
|
||||||
q = add_stat_columns(q)
|
q = add_stat_columns(q)
|
||||||
|
@ -159,7 +154,7 @@ async def load_shouts_by(_, info, options):
|
||||||
|
|
||||||
order_by = options.get("order_by", Shout.publishedAt)
|
order_by = options.get("order_by", Shout.publishedAt)
|
||||||
|
|
||||||
query_order_by = desc(order_by) if options.get('order_by_desc', True) else asc(order_by)
|
query_order_by = desc(order_by) if options.get("order_by_desc", True) else asc(order_by)
|
||||||
offset = options.get("offset", 0)
|
offset = options.get("offset", 0)
|
||||||
limit = options.get("limit", 10)
|
limit = options.get("limit", 10)
|
||||||
|
|
||||||
|
@ -169,13 +164,15 @@ async def load_shouts_by(_, info, options):
|
||||||
with local_session() as session:
|
with local_session() as session:
|
||||||
shouts_map = {}
|
shouts_map = {}
|
||||||
|
|
||||||
for [shout, reacted_stat, commented_stat, rating_stat, last_comment] in session.execute(q).unique():
|
for [shout, reacted_stat, commented_stat, rating_stat, last_comment] in session.execute(
|
||||||
|
q
|
||||||
|
).unique():
|
||||||
shouts.append(shout)
|
shouts.append(shout)
|
||||||
shout.stat = {
|
shout.stat = {
|
||||||
"viewed": shout.views,
|
"viewed": shout.views,
|
||||||
"reacted": reacted_stat,
|
"reacted": reacted_stat,
|
||||||
"commented": commented_stat,
|
"commented": commented_stat,
|
||||||
"rating": rating_stat
|
"rating": rating_stat,
|
||||||
}
|
}
|
||||||
shouts_map[shout.id] = shout
|
shouts_map[shout.id] = shout
|
||||||
|
|
||||||
|
@ -188,11 +185,13 @@ async def get_drafts(_, info):
|
||||||
auth: AuthCredentials = info.context["request"].auth
|
auth: AuthCredentials = info.context["request"].auth
|
||||||
user_id = auth.user_id
|
user_id = auth.user_id
|
||||||
|
|
||||||
q = select(Shout).options(
|
q = (
|
||||||
joinedload(Shout.authors),
|
select(Shout)
|
||||||
joinedload(Shout.topics),
|
.options(
|
||||||
).where(
|
joinedload(Shout.authors),
|
||||||
and_(Shout.deletedAt.is_(None), Shout.createdBy == user_id)
|
joinedload(Shout.topics),
|
||||||
|
)
|
||||||
|
.where(and_(Shout.deletedAt.is_(None), Shout.createdBy == user_id))
|
||||||
)
|
)
|
||||||
|
|
||||||
q = q.group_by(Shout.id)
|
q = q.group_by(Shout.id)
|
||||||
|
@ -211,24 +210,22 @@ async def get_my_feed(_, info, options):
|
||||||
auth: AuthCredentials = info.context["request"].auth
|
auth: AuthCredentials = info.context["request"].auth
|
||||||
user_id = auth.user_id
|
user_id = auth.user_id
|
||||||
|
|
||||||
subquery = select(Shout.id).join(
|
subquery = (
|
||||||
ShoutAuthor
|
select(Shout.id)
|
||||||
).join(
|
.join(ShoutAuthor)
|
||||||
AuthorFollower, AuthorFollower.follower == user_id
|
.join(AuthorFollower, AuthorFollower.follower == user_id)
|
||||||
).join(
|
.join(ShoutTopic)
|
||||||
ShoutTopic
|
.join(TopicFollower, TopicFollower.follower == user_id)
|
||||||
).join(
|
|
||||||
TopicFollower, TopicFollower.follower == user_id
|
|
||||||
)
|
)
|
||||||
|
|
||||||
q = select(Shout).options(
|
q = (
|
||||||
joinedload(Shout.authors),
|
select(Shout)
|
||||||
joinedload(Shout.topics),
|
.options(
|
||||||
).where(
|
joinedload(Shout.authors),
|
||||||
and_(
|
joinedload(Shout.topics),
|
||||||
Shout.publishedAt.is_not(None),
|
)
|
||||||
Shout.deletedAt.is_(None),
|
.where(
|
||||||
Shout.id.in_(subquery)
|
and_(Shout.publishedAt.is_not(None), Shout.deletedAt.is_(None), Shout.id.in_(subquery))
|
||||||
)
|
)
|
||||||
)
|
)
|
||||||
|
|
||||||
|
@ -237,7 +234,7 @@ async def get_my_feed(_, info, options):
|
||||||
|
|
||||||
order_by = options.get("order_by", Shout.publishedAt)
|
order_by = options.get("order_by", Shout.publishedAt)
|
||||||
|
|
||||||
query_order_by = desc(order_by) if options.get('order_by_desc', True) else asc(order_by)
|
query_order_by = desc(order_by) if options.get("order_by_desc", True) else asc(order_by)
|
||||||
offset = options.get("offset", 0)
|
offset = options.get("offset", 0)
|
||||||
limit = options.get("limit", 10)
|
limit = options.get("limit", 10)
|
||||||
|
|
||||||
|
@ -246,13 +243,15 @@ async def get_my_feed(_, info, options):
|
||||||
shouts = []
|
shouts = []
|
||||||
with local_session() as session:
|
with local_session() as session:
|
||||||
shouts_map = {}
|
shouts_map = {}
|
||||||
for [shout, reacted_stat, commented_stat, rating_stat, last_comment] in session.execute(q).unique():
|
for [shout, reacted_stat, commented_stat, rating_stat, last_comment] in session.execute(
|
||||||
|
q
|
||||||
|
).unique():
|
||||||
shouts.append(shout)
|
shouts.append(shout)
|
||||||
shout.stat = {
|
shout.stat = {
|
||||||
"viewed": shout.views,
|
"viewed": shout.views,
|
||||||
"reacted": reacted_stat,
|
"reacted": reacted_stat,
|
||||||
"commented": commented_stat,
|
"commented": commented_stat,
|
||||||
"rating": rating_stat
|
"rating": rating_stat,
|
||||||
}
|
}
|
||||||
shouts_map[shout.id] = shout
|
shouts_map[shout.id] = shout
|
||||||
|
|
||||||
|
|
|
@ -1,6 +1,7 @@
|
||||||
from typing import List
|
|
||||||
from datetime import datetime, timedelta, timezone
|
from datetime import datetime, timedelta, timezone
|
||||||
from sqlalchemy import and_, func, distinct, select, literal
|
from typing import List
|
||||||
|
|
||||||
|
from sqlalchemy import and_, distinct, func, literal, select
|
||||||
from sqlalchemy.orm import aliased, joinedload
|
from sqlalchemy.orm import aliased, joinedload
|
||||||
|
|
||||||
from auth.authenticate import login_required
|
from auth.authenticate import login_required
|
||||||
|
@ -21,27 +22,27 @@ def add_author_stat_columns(q):
|
||||||
# user_rating_aliased = aliased(UserRating)
|
# user_rating_aliased = aliased(UserRating)
|
||||||
|
|
||||||
q = q.outerjoin(shout_author_aliased).add_columns(
|
q = q.outerjoin(shout_author_aliased).add_columns(
|
||||||
func.count(distinct(shout_author_aliased.shout)).label('shouts_stat')
|
func.count(distinct(shout_author_aliased.shout)).label("shouts_stat")
|
||||||
)
|
)
|
||||||
q = q.outerjoin(author_followers, author_followers.author == User.id).add_columns(
|
q = q.outerjoin(author_followers, author_followers.author == User.id).add_columns(
|
||||||
func.count(distinct(author_followers.follower)).label('followers_stat')
|
func.count(distinct(author_followers.follower)).label("followers_stat")
|
||||||
)
|
)
|
||||||
|
|
||||||
q = q.outerjoin(author_following, author_following.follower == User.id).add_columns(
|
q = q.outerjoin(author_following, author_following.follower == User.id).add_columns(
|
||||||
func.count(distinct(author_following.author)).label('followings_stat')
|
func.count(distinct(author_following.author)).label("followings_stat")
|
||||||
)
|
)
|
||||||
|
|
||||||
q = q.add_columns(literal(0).label('rating_stat'))
|
q = q.add_columns(literal(0).label("rating_stat"))
|
||||||
# FIXME
|
# FIXME
|
||||||
# q = q.outerjoin(user_rating_aliased, user_rating_aliased.user == User.id).add_columns(
|
# q = q.outerjoin(user_rating_aliased, user_rating_aliased.user == User.id).add_columns(
|
||||||
# # TODO: check
|
# # TODO: check
|
||||||
# func.sum(user_rating_aliased.value).label('rating_stat')
|
# func.sum(user_rating_aliased.value).label('rating_stat')
|
||||||
# )
|
# )
|
||||||
|
|
||||||
q = q.add_columns(literal(0).label('commented_stat'))
|
q = q.add_columns(literal(0).label("commented_stat"))
|
||||||
# q = q.outerjoin(Reaction, and_(Reaction.createdBy == User.id, Reaction.body.is_not(None))).add_columns(
|
# q = q.outerjoin(
|
||||||
# func.count(distinct(Reaction.id)).label('commented_stat')
|
# Reaction, and_(Reaction.createdBy == User.id, Reaction.body.is_not(None))
|
||||||
# )
|
# ).add_columns(func.count(distinct(Reaction.id)).label("commented_stat"))
|
||||||
|
|
||||||
q = q.group_by(User.id)
|
q = q.group_by(User.id)
|
||||||
|
|
||||||
|
@ -55,7 +56,7 @@ def add_stat(author, stat_columns):
|
||||||
"followers": followers_stat,
|
"followers": followers_stat,
|
||||||
"followings": followings_stat,
|
"followings": followings_stat,
|
||||||
"rating": rating_stat,
|
"rating": rating_stat,
|
||||||
"commented": commented_stat
|
"commented": commented_stat,
|
||||||
}
|
}
|
||||||
|
|
||||||
return author
|
return author
|
||||||
|
@ -119,10 +120,10 @@ async def user_followers(_, _info, slug) -> List[User]:
|
||||||
q = add_author_stat_columns(q)
|
q = add_author_stat_columns(q)
|
||||||
|
|
||||||
aliased_user = aliased(User)
|
aliased_user = aliased(User)
|
||||||
q = q.join(AuthorFollower, AuthorFollower.follower == User.id).join(
|
q = (
|
||||||
aliased_user, aliased_user.id == AuthorFollower.author
|
q.join(AuthorFollower, AuthorFollower.follower == User.id)
|
||||||
).where(
|
.join(aliased_user, aliased_user.id == AuthorFollower.author)
|
||||||
aliased_user.slug == slug
|
.where(aliased_user.slug == slug)
|
||||||
)
|
)
|
||||||
|
|
||||||
return get_authors_from_query(q)
|
return get_authors_from_query(q)
|
||||||
|
@ -150,15 +151,10 @@ async def update_profile(_, info, profile):
|
||||||
with local_session() as session:
|
with local_session() as session:
|
||||||
user = session.query(User).filter(User.id == user_id).one()
|
user = session.query(User).filter(User.id == user_id).one()
|
||||||
if not user:
|
if not user:
|
||||||
return {
|
return {"error": "canoot find user"}
|
||||||
"error": "canoot find user"
|
|
||||||
}
|
|
||||||
user.update(profile)
|
user.update(profile)
|
||||||
session.commit()
|
session.commit()
|
||||||
return {
|
return {"error": None, "author": user}
|
||||||
"error": None,
|
|
||||||
"author": user
|
|
||||||
}
|
|
||||||
|
|
||||||
|
|
||||||
@mutation.field("rateUser")
|
@mutation.field("rateUser")
|
||||||
|
@ -192,7 +188,8 @@ def author_follow(user_id, slug):
|
||||||
session.add(af)
|
session.add(af)
|
||||||
session.commit()
|
session.commit()
|
||||||
return True
|
return True
|
||||||
except:
|
except Exception as e:
|
||||||
|
print(e)
|
||||||
return False
|
return False
|
||||||
|
|
||||||
|
|
||||||
|
@ -200,13 +197,10 @@ def author_follow(user_id, slug):
|
||||||
def author_unfollow(user_id, slug):
|
def author_unfollow(user_id, slug):
|
||||||
with local_session() as session:
|
with local_session() as session:
|
||||||
flw = (
|
flw = (
|
||||||
session.query(
|
session.query(AuthorFollower)
|
||||||
AuthorFollower
|
.join(User, User.id == AuthorFollower.author)
|
||||||
).join(User, User.id == AuthorFollower.author).filter(
|
.filter(and_(AuthorFollower.follower == user_id, User.slug == slug))
|
||||||
and_(
|
.first()
|
||||||
AuthorFollower.follower == user_id, User.slug == slug
|
|
||||||
)
|
|
||||||
).first()
|
|
||||||
)
|
)
|
||||||
if flw:
|
if flw:
|
||||||
session.delete(flw)
|
session.delete(flw)
|
||||||
|
@ -232,12 +226,11 @@ async def get_author(_, _info, slug):
|
||||||
[author] = get_authors_from_query(q)
|
[author] = get_authors_from_query(q)
|
||||||
|
|
||||||
with local_session() as session:
|
with local_session() as session:
|
||||||
comments_count = session.query(Reaction).where(
|
comments_count = (
|
||||||
and_(
|
session.query(Reaction)
|
||||||
Reaction.createdBy == author.id,
|
.where(and_(Reaction.createdBy == author.id, Reaction.kind == ReactionKind.COMMENT))
|
||||||
Reaction.kind == ReactionKind.COMMENT
|
.count()
|
||||||
)
|
)
|
||||||
).count()
|
|
||||||
author.stat["commented"] = comments_count
|
author.stat["commented"] = comments_count
|
||||||
|
|
||||||
return author
|
return author
|
||||||
|
@ -260,9 +253,7 @@ async def load_authors_by(_, info, by, limit, offset):
|
||||||
days_before = datetime.now(tz=timezone.utc) - timedelta(days=by["createdAt"])
|
days_before = datetime.now(tz=timezone.utc) - timedelta(days=by["createdAt"])
|
||||||
q = q.filter(User.createdAt > days_before)
|
q = q.filter(User.createdAt > days_before)
|
||||||
|
|
||||||
q = q.order_by(
|
q = q.order_by(by.get("order", User.createdAt)).limit(limit).offset(offset)
|
||||||
by.get("order", User.createdAt)
|
|
||||||
).limit(limit).offset(offset)
|
|
||||||
|
|
||||||
return get_authors_from_query(q)
|
return get_authors_from_query(q)
|
||||||
|
|
||||||
|
@ -273,13 +264,13 @@ async def load_my_subscriptions(_, info):
|
||||||
auth = info.context["request"].auth
|
auth = info.context["request"].auth
|
||||||
user_id = auth.user_id
|
user_id = auth.user_id
|
||||||
|
|
||||||
authors_query = select(User).join(AuthorFollower, AuthorFollower.author == User.id).where(
|
authors_query = (
|
||||||
AuthorFollower.follower == user_id
|
select(User)
|
||||||
|
.join(AuthorFollower, AuthorFollower.author == User.id)
|
||||||
|
.where(AuthorFollower.follower == user_id)
|
||||||
)
|
)
|
||||||
|
|
||||||
topics_query = select(Topic).join(TopicFollower).where(
|
topics_query = select(Topic).join(TopicFollower).where(TopicFollower.follower == user_id)
|
||||||
TopicFollower.follower == user_id
|
|
||||||
)
|
|
||||||
|
|
||||||
topics = []
|
topics = []
|
||||||
authors = []
|
authors = []
|
||||||
|
@ -291,7 +282,4 @@ async def load_my_subscriptions(_, info):
|
||||||
for [topic] in session.execute(topics_query):
|
for [topic] in session.execute(topics_query):
|
||||||
topics.append(topic)
|
topics.append(topic)
|
||||||
|
|
||||||
return {
|
return {"topics": topics, "authors": authors}
|
||||||
"topics": topics,
|
|
||||||
"authors": authors
|
|
||||||
}
|
|
||||||
|
|
|
@ -1,5 +1,6 @@
|
||||||
from datetime import datetime, timedelta, timezone
|
from datetime import datetime, timedelta, timezone
|
||||||
from sqlalchemy import and_, asc, desc, select, text, func, case
|
|
||||||
|
from sqlalchemy import and_, asc, case, desc, func, select, text
|
||||||
from sqlalchemy.orm import aliased
|
from sqlalchemy.orm import aliased
|
||||||
|
|
||||||
from auth.authenticate import login_required
|
from auth.authenticate import login_required
|
||||||
|
@ -17,26 +18,22 @@ def add_reaction_stat_columns(q):
|
||||||
aliased_reaction = aliased(Reaction)
|
aliased_reaction = aliased(Reaction)
|
||||||
|
|
||||||
q = q.outerjoin(aliased_reaction, Reaction.id == aliased_reaction.replyTo).add_columns(
|
q = q.outerjoin(aliased_reaction, Reaction.id == aliased_reaction.replyTo).add_columns(
|
||||||
func.sum(
|
func.sum(aliased_reaction.id).label("reacted_stat"),
|
||||||
aliased_reaction.id
|
func.sum(case((aliased_reaction.body.is_not(None), 1), else_=0)).label("commented_stat"),
|
||||||
).label('reacted_stat'),
|
|
||||||
func.sum(
|
func.sum(
|
||||||
case(
|
case(
|
||||||
(aliased_reaction.body.is_not(None), 1),
|
(aliased_reaction.kind == ReactionKind.AGREE, 1),
|
||||||
else_=0
|
(aliased_reaction.kind == ReactionKind.DISAGREE, -1),
|
||||||
|
(aliased_reaction.kind == ReactionKind.PROOF, 1),
|
||||||
|
(aliased_reaction.kind == ReactionKind.DISPROOF, -1),
|
||||||
|
(aliased_reaction.kind == ReactionKind.ACCEPT, 1),
|
||||||
|
(aliased_reaction.kind == ReactionKind.REJECT, -1),
|
||||||
|
(aliased_reaction.kind == ReactionKind.LIKE, 1),
|
||||||
|
(aliased_reaction.kind == ReactionKind.DISLIKE, -1),
|
||||||
|
else_=0,
|
||||||
)
|
)
|
||||||
).label('commented_stat'),
|
).label("rating_stat"),
|
||||||
func.sum(case(
|
)
|
||||||
(aliased_reaction.kind == ReactionKind.AGREE, 1),
|
|
||||||
(aliased_reaction.kind == ReactionKind.DISAGREE, -1),
|
|
||||||
(aliased_reaction.kind == ReactionKind.PROOF, 1),
|
|
||||||
(aliased_reaction.kind == ReactionKind.DISPROOF, -1),
|
|
||||||
(aliased_reaction.kind == ReactionKind.ACCEPT, 1),
|
|
||||||
(aliased_reaction.kind == ReactionKind.REJECT, -1),
|
|
||||||
(aliased_reaction.kind == ReactionKind.LIKE, 1),
|
|
||||||
(aliased_reaction.kind == ReactionKind.DISLIKE, -1),
|
|
||||||
else_=0)
|
|
||||||
).label('rating_stat'))
|
|
||||||
|
|
||||||
return q
|
return q
|
||||||
|
|
||||||
|
@ -47,22 +44,25 @@ def reactions_follow(user_id, shout_id: int, auto=False):
|
||||||
shout = session.query(Shout).where(Shout.id == shout_id).one()
|
shout = session.query(Shout).where(Shout.id == shout_id).one()
|
||||||
|
|
||||||
following = (
|
following = (
|
||||||
session.query(ShoutReactionsFollower).where(and_(
|
session.query(ShoutReactionsFollower)
|
||||||
ShoutReactionsFollower.follower == user_id,
|
.where(
|
||||||
ShoutReactionsFollower.shout == shout.id,
|
and_(
|
||||||
)).first()
|
ShoutReactionsFollower.follower == user_id,
|
||||||
|
ShoutReactionsFollower.shout == shout.id,
|
||||||
|
)
|
||||||
|
)
|
||||||
|
.first()
|
||||||
)
|
)
|
||||||
|
|
||||||
if not following:
|
if not following:
|
||||||
following = ShoutReactionsFollower.create(
|
following = ShoutReactionsFollower.create(
|
||||||
follower=user_id,
|
follower=user_id, shout=shout.id, auto=auto
|
||||||
shout=shout.id,
|
|
||||||
auto=auto
|
|
||||||
)
|
)
|
||||||
session.add(following)
|
session.add(following)
|
||||||
session.commit()
|
session.commit()
|
||||||
return True
|
return True
|
||||||
except:
|
except Exception as e:
|
||||||
|
print(e)
|
||||||
return False
|
return False
|
||||||
|
|
||||||
|
|
||||||
|
@ -72,46 +72,52 @@ def reactions_unfollow(user_id: int, shout_id: int):
|
||||||
shout = session.query(Shout).where(Shout.id == shout_id).one()
|
shout = session.query(Shout).where(Shout.id == shout_id).one()
|
||||||
|
|
||||||
following = (
|
following = (
|
||||||
session.query(ShoutReactionsFollower).where(and_(
|
session.query(ShoutReactionsFollower)
|
||||||
ShoutReactionsFollower.follower == user_id,
|
.where(
|
||||||
ShoutReactionsFollower.shout == shout.id
|
and_(
|
||||||
)).first()
|
ShoutReactionsFollower.follower == user_id,
|
||||||
|
ShoutReactionsFollower.shout == shout.id,
|
||||||
|
)
|
||||||
|
)
|
||||||
|
.first()
|
||||||
)
|
)
|
||||||
|
|
||||||
if following:
|
if following:
|
||||||
session.delete(following)
|
session.delete(following)
|
||||||
session.commit()
|
session.commit()
|
||||||
return True
|
return True
|
||||||
except:
|
except Exception as e:
|
||||||
|
print(e)
|
||||||
pass
|
pass
|
||||||
return False
|
return False
|
||||||
|
|
||||||
|
|
||||||
def is_published_author(session, user_id):
|
def is_published_author(session, user_id):
|
||||||
''' checks if user has at least one publication '''
|
"""checks if user has at least one publication"""
|
||||||
return session.query(
|
return (
|
||||||
Shout
|
session.query(Shout)
|
||||||
).where(
|
.where(Shout.authors.contains(user_id))
|
||||||
Shout.authors.contains(user_id)
|
.filter(and_(Shout.publishedAt.is_not(None), Shout.deletedAt.is_(None)))
|
||||||
).filter(
|
.count()
|
||||||
and_(
|
> 0
|
||||||
Shout.publishedAt.is_not(None),
|
)
|
||||||
Shout.deletedAt.is_(None)
|
|
||||||
)
|
|
||||||
).count() > 0
|
|
||||||
|
|
||||||
|
|
||||||
def check_to_publish(session, user_id, reaction):
|
def check_to_publish(session, user_id, reaction):
|
||||||
''' set shout to public if publicated approvers amount > 4 '''
|
"""set shout to public if publicated approvers amount > 4"""
|
||||||
if not reaction.replyTo and reaction.kind in [
|
if not reaction.replyTo and reaction.kind in [
|
||||||
ReactionKind.ACCEPT,
|
ReactionKind.ACCEPT,
|
||||||
ReactionKind.LIKE,
|
ReactionKind.LIKE,
|
||||||
ReactionKind.PROOF
|
ReactionKind.PROOF,
|
||||||
]:
|
]:
|
||||||
if is_published_author(user_id):
|
if is_published_author(user_id):
|
||||||
# now count how many approvers are voted already
|
# now count how many approvers are voted already
|
||||||
approvers_reactions = session.query(Reaction).where(Reaction.shout == reaction.shout).all()
|
approvers_reactions = (
|
||||||
approvers = [user_id, ]
|
session.query(Reaction).where(Reaction.shout == reaction.shout).all()
|
||||||
|
)
|
||||||
|
approvers = [
|
||||||
|
user_id,
|
||||||
|
]
|
||||||
for ar in approvers_reactions:
|
for ar in approvers_reactions:
|
||||||
a = ar.createdBy
|
a = ar.createdBy
|
||||||
if is_published_author(session, a):
|
if is_published_author(session, a):
|
||||||
|
@ -122,21 +128,17 @@ def check_to_publish(session, user_id, reaction):
|
||||||
|
|
||||||
|
|
||||||
def check_to_hide(session, user_id, reaction):
|
def check_to_hide(session, user_id, reaction):
|
||||||
''' hides any shout if 20% of reactions are negative '''
|
"""hides any shout if 20% of reactions are negative"""
|
||||||
if not reaction.replyTo and reaction.kind in [
|
if not reaction.replyTo and reaction.kind in [
|
||||||
ReactionKind.REJECT,
|
ReactionKind.REJECT,
|
||||||
ReactionKind.DISLIKE,
|
ReactionKind.DISLIKE,
|
||||||
ReactionKind.DISPROOF
|
ReactionKind.DISPROOF,
|
||||||
]:
|
]:
|
||||||
# if is_published_author(user):
|
# if is_published_author(user):
|
||||||
approvers_reactions = session.query(Reaction).where(Reaction.shout == reaction.shout).all()
|
approvers_reactions = session.query(Reaction).where(Reaction.shout == reaction.shout).all()
|
||||||
rejects = 0
|
rejects = 0
|
||||||
for r in approvers_reactions:
|
for r in approvers_reactions:
|
||||||
if r.kind in [
|
if r.kind in [ReactionKind.REJECT, ReactionKind.DISLIKE, ReactionKind.DISPROOF]:
|
||||||
ReactionKind.REJECT,
|
|
||||||
ReactionKind.DISLIKE,
|
|
||||||
ReactionKind.DISPROOF
|
|
||||||
]:
|
|
||||||
rejects += 1
|
rejects += 1
|
||||||
if len(approvers_reactions) / rejects < 5:
|
if len(approvers_reactions) / rejects < 5:
|
||||||
return True
|
return True
|
||||||
|
@ -146,14 +148,14 @@ def check_to_hide(session, user_id, reaction):
|
||||||
def set_published(session, shout_id):
|
def set_published(session, shout_id):
|
||||||
s = session.query(Shout).where(Shout.id == shout_id).first()
|
s = session.query(Shout).where(Shout.id == shout_id).first()
|
||||||
s.publishedAt = datetime.now(tz=timezone.utc)
|
s.publishedAt = datetime.now(tz=timezone.utc)
|
||||||
s.visibility = text('public')
|
s.visibility = text("public")
|
||||||
session.add(s)
|
session.add(s)
|
||||||
session.commit()
|
session.commit()
|
||||||
|
|
||||||
|
|
||||||
def set_hidden(session, shout_id):
|
def set_hidden(session, shout_id):
|
||||||
s = session.query(Shout).where(Shout.id == shout_id).first()
|
s = session.query(Shout).where(Shout.id == shout_id).first()
|
||||||
s.visibility = text('community')
|
s.visibility = text("community")
|
||||||
session.add(s)
|
session.add(s)
|
||||||
session.commit()
|
session.commit()
|
||||||
|
|
||||||
|
@ -162,37 +164,46 @@ def set_hidden(session, shout_id):
|
||||||
@login_required
|
@login_required
|
||||||
async def create_reaction(_, info, reaction):
|
async def create_reaction(_, info, reaction):
|
||||||
auth: AuthCredentials = info.context["request"].auth
|
auth: AuthCredentials = info.context["request"].auth
|
||||||
reaction['createdBy'] = auth.user_id
|
reaction["createdBy"] = auth.user_id
|
||||||
rdict = {}
|
rdict = {}
|
||||||
with local_session() as session:
|
with local_session() as session:
|
||||||
shout = session.query(Shout).where(Shout.id == reaction["shout"]).one()
|
shout = session.query(Shout).where(Shout.id == reaction["shout"]).one()
|
||||||
author = session.query(User).where(User.id == auth.user_id).one()
|
author = session.query(User).where(User.id == auth.user_id).one()
|
||||||
|
|
||||||
if reaction["kind"] in [
|
if reaction["kind"] in [ReactionKind.DISLIKE.name, ReactionKind.LIKE.name]:
|
||||||
ReactionKind.DISLIKE.name,
|
existing_reaction = (
|
||||||
ReactionKind.LIKE.name
|
session.query(Reaction)
|
||||||
]:
|
.where(
|
||||||
existing_reaction = session.query(Reaction).where(
|
and_(
|
||||||
and_(
|
Reaction.shout == reaction["shout"],
|
||||||
Reaction.shout == reaction["shout"],
|
Reaction.createdBy == auth.user_id,
|
||||||
Reaction.createdBy == auth.user_id,
|
Reaction.kind == reaction["kind"],
|
||||||
Reaction.kind == reaction["kind"],
|
Reaction.replyTo == reaction.get("replyTo"),
|
||||||
Reaction.replyTo == reaction.get("replyTo")
|
)
|
||||||
)
|
)
|
||||||
).first()
|
.first()
|
||||||
|
)
|
||||||
|
|
||||||
if existing_reaction is not None:
|
if existing_reaction is not None:
|
||||||
raise OperationNotAllowed("You can't vote twice")
|
raise OperationNotAllowed("You can't vote twice")
|
||||||
|
|
||||||
opposite_reaction_kind = ReactionKind.DISLIKE if reaction["kind"] == ReactionKind.LIKE.name else ReactionKind.LIKE
|
opposite_reaction_kind = (
|
||||||
opposite_reaction = session.query(Reaction).where(
|
ReactionKind.DISLIKE
|
||||||
|
if reaction["kind"] == ReactionKind.LIKE.name
|
||||||
|
else ReactionKind.LIKE
|
||||||
|
)
|
||||||
|
opposite_reaction = (
|
||||||
|
session.query(Reaction)
|
||||||
|
.where(
|
||||||
and_(
|
and_(
|
||||||
Reaction.shout == reaction["shout"],
|
Reaction.shout == reaction["shout"],
|
||||||
Reaction.createdBy == auth.user_id,
|
Reaction.createdBy == auth.user_id,
|
||||||
Reaction.kind == opposite_reaction_kind,
|
Reaction.kind == opposite_reaction_kind,
|
||||||
Reaction.replyTo == reaction.get("replyTo")
|
Reaction.replyTo == reaction.get("replyTo"),
|
||||||
)
|
)
|
||||||
).first()
|
)
|
||||||
|
.first()
|
||||||
|
)
|
||||||
|
|
||||||
if opposite_reaction is not None:
|
if opposite_reaction is not None:
|
||||||
session.delete(opposite_reaction)
|
session.delete(opposite_reaction)
|
||||||
|
@ -221,8 +232,8 @@ async def create_reaction(_, info, reaction):
|
||||||
await notification_service.handle_new_reaction(r.id)
|
await notification_service.handle_new_reaction(r.id)
|
||||||
|
|
||||||
rdict = r.dict()
|
rdict = r.dict()
|
||||||
rdict['shout'] = shout.dict()
|
rdict["shout"] = shout.dict()
|
||||||
rdict['createdBy'] = author.dict()
|
rdict["createdBy"] = author.dict()
|
||||||
|
|
||||||
# self-regulation mechanics
|
# self-regulation mechanics
|
||||||
if check_to_hide(session, auth.user_id, r):
|
if check_to_hide(session, auth.user_id, r):
|
||||||
|
@ -235,11 +246,7 @@ async def create_reaction(_, info, reaction):
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
print(f"[resolvers.reactions] error on reactions autofollowing: {e}")
|
print(f"[resolvers.reactions] error on reactions autofollowing: {e}")
|
||||||
|
|
||||||
rdict['stat'] = {
|
rdict["stat"] = {"commented": 0, "reacted": 0, "rating": 0}
|
||||||
"commented": 0,
|
|
||||||
"reacted": 0,
|
|
||||||
"rating": 0
|
|
||||||
}
|
|
||||||
return {"reaction": rdict}
|
return {"reaction": rdict}
|
||||||
|
|
||||||
|
|
||||||
|
@ -269,11 +276,7 @@ async def update_reaction(_, info, id, reaction={}):
|
||||||
if reaction.get("range"):
|
if reaction.get("range"):
|
||||||
r.range = reaction.get("range")
|
r.range = reaction.get("range")
|
||||||
session.commit()
|
session.commit()
|
||||||
r.stat = {
|
r.stat = {"commented": commented_stat, "reacted": reacted_stat, "rating": rating_stat}
|
||||||
"commented": commented_stat,
|
|
||||||
"reacted": reacted_stat,
|
|
||||||
"rating": rating_stat
|
|
||||||
}
|
|
||||||
|
|
||||||
return {"reaction": r}
|
return {"reaction": r}
|
||||||
|
|
||||||
|
@ -290,17 +293,12 @@ async def delete_reaction(_, info, id):
|
||||||
if r.createdBy != auth.user_id:
|
if r.createdBy != auth.user_id:
|
||||||
return {"error": "access denied"}
|
return {"error": "access denied"}
|
||||||
|
|
||||||
if r.kind in [
|
if r.kind in [ReactionKind.LIKE, ReactionKind.DISLIKE]:
|
||||||
ReactionKind.LIKE,
|
|
||||||
ReactionKind.DISLIKE
|
|
||||||
]:
|
|
||||||
session.delete(r)
|
session.delete(r)
|
||||||
else:
|
else:
|
||||||
r.deletedAt = datetime.now(tz=timezone.utc)
|
r.deletedAt = datetime.now(tz=timezone.utc)
|
||||||
session.commit()
|
session.commit()
|
||||||
return {
|
return {"reaction": r}
|
||||||
"reaction": r
|
|
||||||
}
|
|
||||||
|
|
||||||
|
|
||||||
@query.field("loadReactionsBy")
|
@query.field("loadReactionsBy")
|
||||||
|
@ -321,12 +319,10 @@ async def load_reactions_by(_, _info, by, limit=50, offset=0):
|
||||||
:return: Reaction[]
|
:return: Reaction[]
|
||||||
"""
|
"""
|
||||||
|
|
||||||
q = select(
|
q = (
|
||||||
Reaction, User, Shout
|
select(Reaction, User, Shout)
|
||||||
).join(
|
.join(User, Reaction.createdBy == User.id)
|
||||||
User, Reaction.createdBy == User.id
|
.join(Shout, Reaction.shout == Shout.id)
|
||||||
).join(
|
|
||||||
Shout, Reaction.shout == Shout.id
|
|
||||||
)
|
)
|
||||||
|
|
||||||
if by.get("shout"):
|
if by.get("shout"):
|
||||||
|
@ -344,7 +340,7 @@ async def load_reactions_by(_, _info, by, limit=50, offset=0):
|
||||||
if by.get("comment"):
|
if by.get("comment"):
|
||||||
q = q.filter(func.length(Reaction.body) > 0)
|
q = q.filter(func.length(Reaction.body) > 0)
|
||||||
|
|
||||||
if len(by.get('search', '')) > 2:
|
if len(by.get("search", "")) > 2:
|
||||||
q = q.filter(Reaction.body.ilike(f'%{by["body"]}%'))
|
q = q.filter(Reaction.body.ilike(f'%{by["body"]}%'))
|
||||||
|
|
||||||
if by.get("days"):
|
if by.get("days"):
|
||||||
|
@ -352,13 +348,9 @@ async def load_reactions_by(_, _info, by, limit=50, offset=0):
|
||||||
q = q.filter(Reaction.createdAt > after)
|
q = q.filter(Reaction.createdAt > after)
|
||||||
|
|
||||||
order_way = asc if by.get("sort", "").startswith("-") else desc
|
order_way = asc if by.get("sort", "").startswith("-") else desc
|
||||||
order_field = by.get("sort", "").replace('-', '') or Reaction.createdAt
|
order_field = by.get("sort", "").replace("-", "") or Reaction.createdAt
|
||||||
|
|
||||||
q = q.group_by(
|
q = q.group_by(Reaction.id, User.id, Shout.id).order_by(order_way(order_field))
|
||||||
Reaction.id, User.id, Shout.id
|
|
||||||
).order_by(
|
|
||||||
order_way(order_field)
|
|
||||||
)
|
|
||||||
|
|
||||||
q = add_reaction_stat_columns(q)
|
q = add_reaction_stat_columns(q)
|
||||||
|
|
||||||
|
@ -367,13 +359,15 @@ async def load_reactions_by(_, _info, by, limit=50, offset=0):
|
||||||
reactions = []
|
reactions = []
|
||||||
|
|
||||||
with local_session() as session:
|
with local_session() as session:
|
||||||
for [reaction, user, shout, reacted_stat, commented_stat, rating_stat] in session.execute(q):
|
for [reaction, user, shout, reacted_stat, commented_stat, rating_stat] in session.execute(
|
||||||
|
q
|
||||||
|
):
|
||||||
reaction.createdBy = user
|
reaction.createdBy = user
|
||||||
reaction.shout = shout
|
reaction.shout = shout
|
||||||
reaction.stat = {
|
reaction.stat = {
|
||||||
"rating": rating_stat,
|
"rating": rating_stat,
|
||||||
"commented": commented_stat,
|
"commented": commented_stat,
|
||||||
"reacted": reacted_stat
|
"reacted": reacted_stat,
|
||||||
}
|
}
|
||||||
|
|
||||||
reaction.kind = reaction.kind.name
|
reaction.kind = reaction.kind.name
|
||||||
|
|
|
@ -1,24 +1,25 @@
|
||||||
from sqlalchemy import and_, select, distinct, func
|
from sqlalchemy import and_, distinct, func, select
|
||||||
from sqlalchemy.orm import aliased
|
from sqlalchemy.orm import aliased
|
||||||
|
|
||||||
from auth.authenticate import login_required
|
from auth.authenticate import login_required
|
||||||
from base.orm import local_session
|
from base.orm import local_session
|
||||||
from base.resolvers import mutation, query
|
from base.resolvers import mutation, query
|
||||||
from orm.shout import ShoutTopic, ShoutAuthor
|
|
||||||
from orm.topic import Topic, TopicFollower
|
|
||||||
from orm import User
|
from orm import User
|
||||||
|
from orm.shout import ShoutAuthor, ShoutTopic
|
||||||
|
from orm.topic import Topic, TopicFollower
|
||||||
|
|
||||||
|
|
||||||
def add_topic_stat_columns(q):
|
def add_topic_stat_columns(q):
|
||||||
aliased_shout_author = aliased(ShoutAuthor)
|
aliased_shout_author = aliased(ShoutAuthor)
|
||||||
aliased_topic_follower = aliased(TopicFollower)
|
aliased_topic_follower = aliased(TopicFollower)
|
||||||
|
|
||||||
q = q.outerjoin(ShoutTopic, Topic.id == ShoutTopic.topic).add_columns(
|
q = (
|
||||||
func.count(distinct(ShoutTopic.shout)).label('shouts_stat')
|
q.outerjoin(ShoutTopic, Topic.id == ShoutTopic.topic)
|
||||||
).outerjoin(aliased_shout_author, ShoutTopic.shout == aliased_shout_author.shout).add_columns(
|
.add_columns(func.count(distinct(ShoutTopic.shout)).label("shouts_stat"))
|
||||||
func.count(distinct(aliased_shout_author.user)).label('authors_stat')
|
.outerjoin(aliased_shout_author, ShoutTopic.shout == aliased_shout_author.shout)
|
||||||
).outerjoin(aliased_topic_follower).add_columns(
|
.add_columns(func.count(distinct(aliased_shout_author.user)).label("authors_stat"))
|
||||||
func.count(distinct(aliased_topic_follower.follower)).label('followers_stat')
|
.outerjoin(aliased_topic_follower)
|
||||||
|
.add_columns(func.count(distinct(aliased_topic_follower.follower)).label("followers_stat"))
|
||||||
)
|
)
|
||||||
|
|
||||||
q = q.group_by(Topic.id)
|
q = q.group_by(Topic.id)
|
||||||
|
@ -28,11 +29,7 @@ def add_topic_stat_columns(q):
|
||||||
|
|
||||||
def add_stat(topic, stat_columns):
|
def add_stat(topic, stat_columns):
|
||||||
[shouts_stat, authors_stat, followers_stat] = stat_columns
|
[shouts_stat, authors_stat, followers_stat] = stat_columns
|
||||||
topic.stat = {
|
topic.stat = {"shouts": shouts_stat, "authors": authors_stat, "followers": followers_stat}
|
||||||
"shouts": shouts_stat,
|
|
||||||
"authors": authors_stat,
|
|
||||||
"followers": followers_stat
|
|
||||||
}
|
|
||||||
|
|
||||||
return topic
|
return topic
|
||||||
|
|
||||||
|
@ -125,7 +122,8 @@ def topic_follow(user_id, slug):
|
||||||
session.add(following)
|
session.add(following)
|
||||||
session.commit()
|
session.commit()
|
||||||
return True
|
return True
|
||||||
except:
|
except Exception as e:
|
||||||
|
print(e)
|
||||||
return False
|
return False
|
||||||
|
|
||||||
|
|
||||||
|
@ -133,18 +131,17 @@ def topic_unfollow(user_id, slug):
|
||||||
try:
|
try:
|
||||||
with local_session() as session:
|
with local_session() as session:
|
||||||
sub = (
|
sub = (
|
||||||
session.query(TopicFollower).join(Topic).filter(
|
session.query(TopicFollower)
|
||||||
and_(
|
.join(Topic)
|
||||||
TopicFollower.follower == user_id,
|
.filter(and_(TopicFollower.follower == user_id, Topic.slug == slug))
|
||||||
Topic.slug == slug
|
.first()
|
||||||
)
|
|
||||||
).first()
|
|
||||||
)
|
)
|
||||||
if sub:
|
if sub:
|
||||||
session.delete(sub)
|
session.delete(sub)
|
||||||
session.commit()
|
session.commit()
|
||||||
return True
|
return True
|
||||||
except:
|
except Exception as e:
|
||||||
|
print(e)
|
||||||
pass
|
pass
|
||||||
return False
|
return False
|
||||||
|
|
||||||
|
|
1157
schema_types.py
Normal file
1157
schema_types.py
Normal file
File diff suppressed because it is too large
Load Diff
83
server.py
83
server.py
|
@ -1,8 +1,9 @@
|
||||||
import sys
|
|
||||||
import os
|
import os
|
||||||
|
import sys
|
||||||
|
|
||||||
import uvicorn
|
import uvicorn
|
||||||
|
|
||||||
from settings import PORT, DEV_SERVER_PID_FILE_NAME
|
from settings import DEV_SERVER_PID_FILE_NAME, PORT
|
||||||
|
|
||||||
|
|
||||||
def exception_handler(exception_type, exception, traceback, debug_hook=sys.excepthook):
|
def exception_handler(exception_type, exception, traceback, debug_hook=sys.excepthook):
|
||||||
|
@ -10,47 +11,36 @@ def exception_handler(exception_type, exception, traceback, debug_hook=sys.excep
|
||||||
|
|
||||||
|
|
||||||
log_settings = {
|
log_settings = {
|
||||||
'version': 1,
|
"version": 1,
|
||||||
'disable_existing_loggers': True,
|
"disable_existing_loggers": True,
|
||||||
'formatters': {
|
"formatters": {
|
||||||
'default': {
|
"default": {
|
||||||
'()': 'uvicorn.logging.DefaultFormatter',
|
"()": "uvicorn.logging.DefaultFormatter",
|
||||||
'fmt': '%(levelprefix)s %(message)s',
|
"fmt": "%(levelprefix)s %(message)s",
|
||||||
'use_colors': None
|
"use_colors": None,
|
||||||
|
},
|
||||||
|
"access": {
|
||||||
|
"()": "uvicorn.logging.AccessFormatter",
|
||||||
|
"fmt": '%(levelprefix)s %(client_addr)s - "%(request_line)s" %(status_code)s',
|
||||||
},
|
},
|
||||||
'access': {
|
|
||||||
'()': 'uvicorn.logging.AccessFormatter',
|
|
||||||
'fmt': '%(levelprefix)s %(client_addr)s - "%(request_line)s" %(status_code)s'
|
|
||||||
}
|
|
||||||
},
|
},
|
||||||
'handlers': {
|
"handlers": {
|
||||||
'default': {
|
"default": {
|
||||||
'formatter': 'default',
|
"formatter": "default",
|
||||||
'class': 'logging.StreamHandler',
|
"class": "logging.StreamHandler",
|
||||||
'stream': 'ext://sys.stderr'
|
"stream": "ext://sys.stderr",
|
||||||
|
},
|
||||||
|
"access": {
|
||||||
|
"formatter": "access",
|
||||||
|
"class": "logging.StreamHandler",
|
||||||
|
"stream": "ext://sys.stdout",
|
||||||
},
|
},
|
||||||
'access': {
|
|
||||||
'formatter': 'access',
|
|
||||||
'class': 'logging.StreamHandler',
|
|
||||||
'stream': 'ext://sys.stdout'
|
|
||||||
}
|
|
||||||
},
|
},
|
||||||
'loggers': {
|
"loggers": {
|
||||||
'uvicorn': {
|
"uvicorn": {"handlers": ["default"], "level": "INFO"},
|
||||||
'handlers': ['default'],
|
"uvicorn.error": {"level": "INFO", "handlers": ["default"], "propagate": True},
|
||||||
'level': 'INFO'
|
"uvicorn.access": {"handlers": ["access"], "level": "INFO", "propagate": False},
|
||||||
},
|
},
|
||||||
'uvicorn.error': {
|
|
||||||
'level': 'INFO',
|
|
||||||
'handlers': ['default'],
|
|
||||||
'propagate': True
|
|
||||||
},
|
|
||||||
'uvicorn.access': {
|
|
||||||
'handlers': ['access'],
|
|
||||||
'level': 'INFO',
|
|
||||||
'propagate': False
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
local_headers = [
|
local_headers = [
|
||||||
|
@ -58,7 +48,8 @@ local_headers = [
|
||||||
("Access-Control-Allow-Origin", "https://localhost:3000"),
|
("Access-Control-Allow-Origin", "https://localhost:3000"),
|
||||||
(
|
(
|
||||||
"Access-Control-Allow-Headers",
|
"Access-Control-Allow-Headers",
|
||||||
"DNT,User-Agent,X-Requested-With,If-Modified-Since,Cache-Control,Content-Type,Range,Authorization",
|
"DNT,User-Agent,X-Requested-With,If-Modified-Since,"
|
||||||
|
+ " Cache-Control,Content-Type,Range,Authorization",
|
||||||
),
|
),
|
||||||
("Access-Control-Expose-Headers", "Content-Length,Content-Range"),
|
("Access-Control-Expose-Headers", "Content-Length,Content-Range"),
|
||||||
("Access-Control-Allow-Credentials", "true"),
|
("Access-Control-Allow-Credentials", "true"),
|
||||||
|
@ -86,24 +77,20 @@ if __name__ == "__main__":
|
||||||
# log_config=log_settings,
|
# log_config=log_settings,
|
||||||
log_level=None,
|
log_level=None,
|
||||||
access_log=True,
|
access_log=True,
|
||||||
reload=want_reload
|
reload=want_reload,
|
||||||
) # , ssl_keyfile="discours.key", ssl_certfile="discours.crt")
|
) # , ssl_keyfile="discours.key", ssl_certfile="discours.crt")
|
||||||
elif x == "migrate":
|
elif x == "migrate":
|
||||||
from migration import process
|
from migration import process
|
||||||
|
|
||||||
print("MODE: MIGRATE")
|
print("MODE: MIGRATE")
|
||||||
|
|
||||||
process()
|
process()
|
||||||
elif x == "bson":
|
elif x == "bson":
|
||||||
from migration.bson2json import json_tables
|
from migration.bson2json import json_tables
|
||||||
|
|
||||||
print("MODE: BSON")
|
print("MODE: BSON")
|
||||||
|
|
||||||
json_tables()
|
json_tables()
|
||||||
else:
|
else:
|
||||||
sys.excepthook = exception_handler
|
sys.excepthook = exception_handler
|
||||||
uvicorn.run(
|
uvicorn.run("main:app", host="0.0.0.0", port=PORT, proxy_headers=True, server_header=True)
|
||||||
"main:app",
|
|
||||||
host="0.0.0.0",
|
|
||||||
port=PORT,
|
|
||||||
proxy_headers=True,
|
|
||||||
server_header=True
|
|
||||||
)
|
|
||||||
|
|
|
@ -18,12 +18,7 @@ class Following:
|
||||||
|
|
||||||
class FollowingManager:
|
class FollowingManager:
|
||||||
lock = asyncio.Lock()
|
lock = asyncio.Lock()
|
||||||
data = {
|
data = {"author": [], "topic": [], "shout": [], "chat": []}
|
||||||
'author': [],
|
|
||||||
'topic': [],
|
|
||||||
'shout': [],
|
|
||||||
'chat': []
|
|
||||||
}
|
|
||||||
|
|
||||||
@staticmethod
|
@staticmethod
|
||||||
async def register(kind, uid):
|
async def register(kind, uid):
|
||||||
|
@ -39,13 +34,13 @@ class FollowingManager:
|
||||||
async def push(kind, payload):
|
async def push(kind, payload):
|
||||||
try:
|
try:
|
||||||
async with FollowingManager.lock:
|
async with FollowingManager.lock:
|
||||||
if kind == 'chat':
|
if kind == "chat":
|
||||||
for chat in FollowingManager['chat']:
|
for chat in FollowingManager["chat"]:
|
||||||
if payload.message["chatId"] == chat.uid:
|
if payload.message["chatId"] == chat.uid:
|
||||||
chat.queue.put_nowait(payload)
|
chat.queue.put_nowait(payload)
|
||||||
else:
|
else:
|
||||||
for entity in FollowingManager[kind]:
|
for entity in FollowingManager[kind]:
|
||||||
if payload.shout['createdBy'] == entity.uid:
|
if payload.shout["createdBy"] == entity.uid:
|
||||||
entity.queue.put_nowait(payload)
|
entity.queue.put_nowait(payload)
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
print(Exception(e))
|
print(Exception(e))
|
||||||
|
|
|
@ -1,13 +1,13 @@
|
||||||
|
from base.orm import local_session
|
||||||
from services.search import SearchService
|
from services.search import SearchService
|
||||||
from services.stat.viewed import ViewedStorage
|
from services.stat.viewed import ViewedStorage
|
||||||
from base.orm import local_session
|
|
||||||
|
|
||||||
|
|
||||||
async def storages_init():
|
async def storages_init():
|
||||||
with local_session() as session:
|
with local_session() as session:
|
||||||
print('[main] initialize SearchService')
|
print("[main] initialize SearchService")
|
||||||
await SearchService.init(session)
|
await SearchService.init(session)
|
||||||
print('[main] SearchService initialized')
|
print("[main] SearchService initialized")
|
||||||
print('[main] initialize storages')
|
print("[main] initialize storages")
|
||||||
await ViewedStorage.init()
|
await ViewedStorage.init()
|
||||||
print('[main] storages initialized')
|
print("[main] storages initialized")
|
||||||
|
|
|
@ -5,32 +5,24 @@ from datetime import datetime, timezone
|
||||||
from sqlalchemy import and_
|
from sqlalchemy import and_
|
||||||
|
|
||||||
from base.orm import local_session
|
from base.orm import local_session
|
||||||
from orm import Reaction, Shout, Notification, User
|
from orm import Notification, Reaction, Shout, User
|
||||||
from orm.notification import NotificationType
|
from orm.notification import NotificationType
|
||||||
from orm.reaction import ReactionKind
|
from orm.reaction import ReactionKind
|
||||||
from services.notifications.sse import connection_manager
|
from services.notifications.sse import connection_manager
|
||||||
|
|
||||||
|
|
||||||
def shout_to_shout_data(shout):
|
def shout_to_shout_data(shout):
|
||||||
return {
|
return {"title": shout.title, "slug": shout.slug}
|
||||||
"title": shout.title,
|
|
||||||
"slug": shout.slug
|
|
||||||
}
|
|
||||||
|
|
||||||
|
|
||||||
def user_to_user_data(user):
|
def user_to_user_data(user):
|
||||||
return {
|
return {"id": user.id, "name": user.name, "slug": user.slug, "userpic": user.userpic}
|
||||||
"id": user.id,
|
|
||||||
"name": user.name,
|
|
||||||
"slug": user.slug,
|
|
||||||
"userpic": user.userpic
|
|
||||||
}
|
|
||||||
|
|
||||||
|
|
||||||
def update_prev_notification(notification, user, reaction):
|
def update_prev_notification(notification, user, reaction):
|
||||||
notification_data = json.loads(notification.data)
|
notification_data = json.loads(notification.data)
|
||||||
|
|
||||||
notification_data["users"] = [u for u in notification_data["users"] if u['id'] != user.id]
|
notification_data["users"] = [u for u in notification_data["users"] if u["id"] != user.id]
|
||||||
notification_data["users"].append(user_to_user_data(user))
|
notification_data["users"].append(user_to_user_data(user))
|
||||||
|
|
||||||
if notification_data["reactionIds"] is None:
|
if notification_data["reactionIds"] is None:
|
||||||
|
@ -57,34 +49,45 @@ class NewReactionNotificator:
|
||||||
if reaction.kind == ReactionKind.COMMENT:
|
if reaction.kind == ReactionKind.COMMENT:
|
||||||
parent_reaction = None
|
parent_reaction = None
|
||||||
if reaction.replyTo:
|
if reaction.replyTo:
|
||||||
parent_reaction = session.query(Reaction).where(Reaction.id == reaction.replyTo).one()
|
parent_reaction = (
|
||||||
|
session.query(Reaction).where(Reaction.id == reaction.replyTo).one()
|
||||||
|
)
|
||||||
if parent_reaction.createdBy != reaction.createdBy:
|
if parent_reaction.createdBy != reaction.createdBy:
|
||||||
prev_new_reply_notification = session.query(Notification).where(
|
prev_new_reply_notification = (
|
||||||
and_(
|
session.query(Notification)
|
||||||
Notification.user == shout.createdBy,
|
.where(
|
||||||
Notification.type == NotificationType.NEW_REPLY,
|
and_(
|
||||||
Notification.shout == shout.id,
|
Notification.user == shout.createdBy,
|
||||||
Notification.reaction == parent_reaction.id,
|
Notification.type == NotificationType.NEW_REPLY,
|
||||||
Notification.seen == False
|
Notification.shout == shout.id,
|
||||||
|
Notification.reaction == parent_reaction.id,
|
||||||
|
Notification.seen == False, # noqa: E712
|
||||||
|
)
|
||||||
)
|
)
|
||||||
).first()
|
.first()
|
||||||
|
)
|
||||||
|
|
||||||
if prev_new_reply_notification:
|
if prev_new_reply_notification:
|
||||||
update_prev_notification(prev_new_reply_notification, user, reaction)
|
update_prev_notification(prev_new_reply_notification, user, reaction)
|
||||||
else:
|
else:
|
||||||
reply_notification_data = json.dumps({
|
reply_notification_data = json.dumps(
|
||||||
"shout": shout_to_shout_data(shout),
|
{
|
||||||
"users": [user_to_user_data(user)],
|
"shout": shout_to_shout_data(shout),
|
||||||
"reactionIds": [reaction.id]
|
"users": [user_to_user_data(user)],
|
||||||
}, ensure_ascii=False)
|
"reactionIds": [reaction.id],
|
||||||
|
},
|
||||||
|
ensure_ascii=False,
|
||||||
|
)
|
||||||
|
|
||||||
reply_notification = Notification.create(**{
|
reply_notification = Notification.create(
|
||||||
"user": parent_reaction.createdBy,
|
**{
|
||||||
"type": NotificationType.NEW_REPLY,
|
"user": parent_reaction.createdBy,
|
||||||
"shout": shout.id,
|
"type": NotificationType.NEW_REPLY,
|
||||||
"reaction": parent_reaction.id,
|
"shout": shout.id,
|
||||||
"data": reply_notification_data
|
"reaction": parent_reaction.id,
|
||||||
})
|
"data": reply_notification_data,
|
||||||
|
}
|
||||||
|
)
|
||||||
|
|
||||||
session.add(reply_notification)
|
session.add(reply_notification)
|
||||||
|
|
||||||
|
@ -93,30 +96,39 @@ class NewReactionNotificator:
|
||||||
if reaction.createdBy != shout.createdBy and (
|
if reaction.createdBy != shout.createdBy and (
|
||||||
parent_reaction is None or parent_reaction.createdBy != shout.createdBy
|
parent_reaction is None or parent_reaction.createdBy != shout.createdBy
|
||||||
):
|
):
|
||||||
prev_new_comment_notification = session.query(Notification).where(
|
prev_new_comment_notification = (
|
||||||
and_(
|
session.query(Notification)
|
||||||
Notification.user == shout.createdBy,
|
.where(
|
||||||
Notification.type == NotificationType.NEW_COMMENT,
|
and_(
|
||||||
Notification.shout == shout.id,
|
Notification.user == shout.createdBy,
|
||||||
Notification.seen == False
|
Notification.type == NotificationType.NEW_COMMENT,
|
||||||
|
Notification.shout == shout.id,
|
||||||
|
Notification.seen == False, # noqa: E712
|
||||||
|
)
|
||||||
)
|
)
|
||||||
).first()
|
.first()
|
||||||
|
)
|
||||||
|
|
||||||
if prev_new_comment_notification:
|
if prev_new_comment_notification:
|
||||||
update_prev_notification(prev_new_comment_notification, user, reaction)
|
update_prev_notification(prev_new_comment_notification, user, reaction)
|
||||||
else:
|
else:
|
||||||
notification_data_string = json.dumps({
|
notification_data_string = json.dumps(
|
||||||
"shout": shout_to_shout_data(shout),
|
{
|
||||||
"users": [user_to_user_data(user)],
|
"shout": shout_to_shout_data(shout),
|
||||||
"reactionIds": [reaction.id]
|
"users": [user_to_user_data(user)],
|
||||||
}, ensure_ascii=False)
|
"reactionIds": [reaction.id],
|
||||||
|
},
|
||||||
|
ensure_ascii=False,
|
||||||
|
)
|
||||||
|
|
||||||
author_notification = Notification.create(**{
|
author_notification = Notification.create(
|
||||||
"user": shout.createdBy,
|
**{
|
||||||
"type": NotificationType.NEW_COMMENT,
|
"user": shout.createdBy,
|
||||||
"shout": shout.id,
|
"type": NotificationType.NEW_COMMENT,
|
||||||
"data": notification_data_string
|
"shout": shout.id,
|
||||||
})
|
"data": notification_data_string,
|
||||||
|
}
|
||||||
|
)
|
||||||
|
|
||||||
session.add(author_notification)
|
session.add(author_notification)
|
||||||
|
|
||||||
|
@ -142,7 +154,7 @@ class NotificationService:
|
||||||
try:
|
try:
|
||||||
await notificator.run()
|
await notificator.run()
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
print(f'[NotificationService.worker] error: {str(e)}')
|
print(f"[NotificationService.worker] error: {str(e)}")
|
||||||
|
|
||||||
|
|
||||||
notification_service = NotificationService()
|
notification_service = NotificationService()
|
||||||
|
|
|
@ -1,8 +1,8 @@
|
||||||
|
import asyncio
|
||||||
import json
|
import json
|
||||||
|
|
||||||
from sse_starlette.sse import EventSourceResponse
|
from sse_starlette.sse import EventSourceResponse
|
||||||
from starlette.requests import Request
|
from starlette.requests import Request
|
||||||
import asyncio
|
|
||||||
|
|
||||||
|
|
||||||
class ConnectionManager:
|
class ConnectionManager:
|
||||||
|
@ -28,9 +28,7 @@ class ConnectionManager:
|
||||||
return
|
return
|
||||||
|
|
||||||
for connection in self.connections_by_user_id[user_id]:
|
for connection in self.connections_by_user_id[user_id]:
|
||||||
data = {
|
data = {"type": "newNotifications"}
|
||||||
"type": "newNotifications"
|
|
||||||
}
|
|
||||||
data_string = json.dumps(data, ensure_ascii=False)
|
data_string = json.dumps(data, ensure_ascii=False)
|
||||||
await connection.put(data_string)
|
await connection.put(data_string)
|
||||||
|
|
||||||
|
|
|
@ -1,5 +1,7 @@
|
||||||
import asyncio
|
import asyncio
|
||||||
import json
|
import json
|
||||||
|
from typing import List
|
||||||
|
|
||||||
from base.redis import redis
|
from base.redis import redis
|
||||||
from orm.shout import Shout
|
from orm.shout import Shout
|
||||||
from resolvers.zine.load import load_shouts_by
|
from resolvers.zine.load import load_shouts_by
|
||||||
|
@ -7,25 +9,20 @@ from resolvers.zine.load import load_shouts_by
|
||||||
|
|
||||||
class SearchService:
|
class SearchService:
|
||||||
lock = asyncio.Lock()
|
lock = asyncio.Lock()
|
||||||
cache = {}
|
# cache = {}
|
||||||
|
|
||||||
@staticmethod
|
@staticmethod
|
||||||
async def init(session):
|
async def init(session):
|
||||||
async with SearchService.lock:
|
async with SearchService.lock:
|
||||||
print('[search.service] did nothing')
|
print("[search.service] did nothing")
|
||||||
SearchService.cache = {}
|
# SearchService.cache = {}
|
||||||
|
|
||||||
@staticmethod
|
@staticmethod
|
||||||
async def search(text, limit, offset) -> [Shout]:
|
async def search(text, limit, offset) -> List[Shout]:
|
||||||
cached = await redis.execute("GET", text)
|
cached = await redis.execute("GET", text)
|
||||||
if not cached:
|
if not cached:
|
||||||
async with SearchService.lock:
|
async with SearchService.lock:
|
||||||
options = {
|
options = {"title": text, "body": text, "limit": limit, "offset": offset}
|
||||||
"title": text,
|
|
||||||
"body": text,
|
|
||||||
"limit": limit,
|
|
||||||
"offset": offset
|
|
||||||
}
|
|
||||||
payload = await load_shouts_by(None, None, options)
|
payload = await load_shouts_by(None, None, options)
|
||||||
await redis.execute("SET", text, json.dumps(payload))
|
await redis.execute("SET", text, json.dumps(payload))
|
||||||
return payload
|
return payload
|
||||||
|
|
|
@ -1,18 +1,18 @@
|
||||||
import asyncio
|
import asyncio
|
||||||
import time
|
import time
|
||||||
from datetime import timedelta, timezone, datetime
|
from datetime import datetime, timedelta, timezone
|
||||||
from os import environ, path
|
from os import environ, path
|
||||||
from ssl import create_default_context
|
from ssl import create_default_context
|
||||||
|
|
||||||
from gql import Client, gql
|
from gql import Client, gql
|
||||||
from gql.transport.aiohttp import AIOHTTPTransport
|
from gql.transport.aiohttp import AIOHTTPTransport
|
||||||
from sqlalchemy import func
|
|
||||||
|
|
||||||
from base.orm import local_session
|
from base.orm import local_session
|
||||||
from orm import User, Topic
|
from orm import Topic
|
||||||
from orm.shout import ShoutTopic, Shout
|
from orm.shout import Shout, ShoutTopic
|
||||||
|
|
||||||
load_facts = gql("""
|
load_facts = gql(
|
||||||
|
"""
|
||||||
query getDomains {
|
query getDomains {
|
||||||
domains {
|
domains {
|
||||||
id
|
id
|
||||||
|
@ -25,9 +25,11 @@ query getDomains {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
""")
|
"""
|
||||||
|
)
|
||||||
|
|
||||||
load_pages = gql("""
|
load_pages = gql(
|
||||||
|
"""
|
||||||
query getDomains {
|
query getDomains {
|
||||||
domains {
|
domains {
|
||||||
title
|
title
|
||||||
|
@ -41,8 +43,9 @@ query getDomains {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
""")
|
"""
|
||||||
schema_str = open(path.dirname(__file__) + '/ackee.graphql').read()
|
)
|
||||||
|
schema_str = open(path.dirname(__file__) + "/ackee.graphql").read()
|
||||||
token = environ.get("ACKEE_TOKEN", "")
|
token = environ.get("ACKEE_TOKEN", "")
|
||||||
|
|
||||||
|
|
||||||
|
@ -50,10 +53,8 @@ def create_client(headers=None, schema=None):
|
||||||
return Client(
|
return Client(
|
||||||
schema=schema,
|
schema=schema,
|
||||||
transport=AIOHTTPTransport(
|
transport=AIOHTTPTransport(
|
||||||
url="https://ackee.discours.io/api",
|
url="https://ackee.discours.io/api", ssl=create_default_context(), headers=headers
|
||||||
ssl=create_default_context(),
|
),
|
||||||
headers=headers
|
|
||||||
)
|
|
||||||
)
|
)
|
||||||
|
|
||||||
|
|
||||||
|
@ -71,13 +72,13 @@ class ViewedStorage:
|
||||||
|
|
||||||
@staticmethod
|
@staticmethod
|
||||||
async def init():
|
async def init():
|
||||||
""" graphql client connection using permanent token """
|
"""graphql client connection using permanent token"""
|
||||||
self = ViewedStorage
|
self = ViewedStorage
|
||||||
async with self.lock:
|
async with self.lock:
|
||||||
if token:
|
if token:
|
||||||
self.client = create_client({
|
self.client = create_client(
|
||||||
"Authorization": "Bearer %s" % str(token)
|
{"Authorization": "Bearer %s" % str(token)}, schema=schema_str
|
||||||
}, schema=schema_str)
|
)
|
||||||
print("[stat.viewed] * authorized permanentely by ackee.discours.io: %s" % token)
|
print("[stat.viewed] * authorized permanentely by ackee.discours.io: %s" % token)
|
||||||
else:
|
else:
|
||||||
print("[stat.viewed] * please set ACKEE_TOKEN")
|
print("[stat.viewed] * please set ACKEE_TOKEN")
|
||||||
|
@ -85,7 +86,7 @@ class ViewedStorage:
|
||||||
|
|
||||||
@staticmethod
|
@staticmethod
|
||||||
async def update_pages():
|
async def update_pages():
|
||||||
""" query all the pages from ackee sorted by views count """
|
"""query all the pages from ackee sorted by views count"""
|
||||||
print("[stat.viewed] ⎧ updating ackee pages data ---")
|
print("[stat.viewed] ⎧ updating ackee pages data ---")
|
||||||
start = time.time()
|
start = time.time()
|
||||||
self = ViewedStorage
|
self = ViewedStorage
|
||||||
|
@ -96,7 +97,7 @@ class ViewedStorage:
|
||||||
try:
|
try:
|
||||||
for page in self.pages:
|
for page in self.pages:
|
||||||
p = page["value"].split("?")[0]
|
p = page["value"].split("?")[0]
|
||||||
slug = p.split('discours.io/')[-1]
|
slug = p.split("discours.io/")[-1]
|
||||||
shouts[slug] = page["count"]
|
shouts[slug] = page["count"]
|
||||||
for slug in shouts.keys():
|
for slug in shouts.keys():
|
||||||
await ViewedStorage.increment(slug, shouts[slug])
|
await ViewedStorage.increment(slug, shouts[slug])
|
||||||
|
@ -118,7 +119,7 @@ class ViewedStorage:
|
||||||
# unused yet
|
# unused yet
|
||||||
@staticmethod
|
@staticmethod
|
||||||
async def get_shout(shout_slug):
|
async def get_shout(shout_slug):
|
||||||
""" getting shout views metric by slug """
|
"""getting shout views metric by slug"""
|
||||||
self = ViewedStorage
|
self = ViewedStorage
|
||||||
async with self.lock:
|
async with self.lock:
|
||||||
shout_views = self.by_shouts.get(shout_slug)
|
shout_views = self.by_shouts.get(shout_slug)
|
||||||
|
@ -136,7 +137,7 @@ class ViewedStorage:
|
||||||
|
|
||||||
@staticmethod
|
@staticmethod
|
||||||
async def get_topic(topic_slug):
|
async def get_topic(topic_slug):
|
||||||
""" getting topic views value summed """
|
"""getting topic views value summed"""
|
||||||
self = ViewedStorage
|
self = ViewedStorage
|
||||||
topic_views = 0
|
topic_views = 0
|
||||||
async with self.lock:
|
async with self.lock:
|
||||||
|
@ -146,24 +147,28 @@ class ViewedStorage:
|
||||||
|
|
||||||
@staticmethod
|
@staticmethod
|
||||||
def update_topics(session, shout_slug):
|
def update_topics(session, shout_slug):
|
||||||
""" updates topics counters by shout slug """
|
"""updates topics counters by shout slug"""
|
||||||
self = ViewedStorage
|
self = ViewedStorage
|
||||||
for [shout_topic, topic] in session.query(ShoutTopic, Topic).join(Topic).join(Shout).where(
|
for [shout_topic, topic] in (
|
||||||
Shout.slug == shout_slug
|
session.query(ShoutTopic, Topic)
|
||||||
).all():
|
.join(Topic)
|
||||||
|
.join(Shout)
|
||||||
|
.where(Shout.slug == shout_slug)
|
||||||
|
.all()
|
||||||
|
):
|
||||||
if not self.by_topics.get(topic.slug):
|
if not self.by_topics.get(topic.slug):
|
||||||
self.by_topics[topic.slug] = {}
|
self.by_topics[topic.slug] = {}
|
||||||
self.by_topics[topic.slug][shout_slug] = self.by_shouts[shout_slug]
|
self.by_topics[topic.slug][shout_slug] = self.by_shouts[shout_slug]
|
||||||
|
|
||||||
@staticmethod
|
@staticmethod
|
||||||
async def increment(shout_slug, amount=1, viewer='ackee'):
|
async def increment(shout_slug, amount=1, viewer="ackee"):
|
||||||
""" the only way to change views counter """
|
"""the only way to change views counter"""
|
||||||
self = ViewedStorage
|
self = ViewedStorage
|
||||||
async with self.lock:
|
async with self.lock:
|
||||||
# TODO optimize, currenty we execute 1 DB transaction per shout
|
# TODO optimize, currenty we execute 1 DB transaction per shout
|
||||||
with local_session() as session:
|
with local_session() as session:
|
||||||
shout = session.query(Shout).where(Shout.slug == shout_slug).one()
|
shout = session.query(Shout).where(Shout.slug == shout_slug).one()
|
||||||
if viewer == 'old-discours':
|
if viewer == "old-discours":
|
||||||
# this is needed for old db migration
|
# this is needed for old db migration
|
||||||
if shout.viewsOld == amount:
|
if shout.viewsOld == amount:
|
||||||
print(f"viewsOld amount: {amount}")
|
print(f"viewsOld amount: {amount}")
|
||||||
|
@ -185,7 +190,7 @@ class ViewedStorage:
|
||||||
|
|
||||||
@staticmethod
|
@staticmethod
|
||||||
async def worker():
|
async def worker():
|
||||||
""" async task worker """
|
"""async task worker"""
|
||||||
failed = 0
|
failed = 0
|
||||||
self = ViewedStorage
|
self = ViewedStorage
|
||||||
if self.disabled:
|
if self.disabled:
|
||||||
|
@ -205,9 +210,10 @@ class ViewedStorage:
|
||||||
if failed == 0:
|
if failed == 0:
|
||||||
when = datetime.now(timezone.utc) + timedelta(seconds=self.period)
|
when = datetime.now(timezone.utc) + timedelta(seconds=self.period)
|
||||||
t = format(when.astimezone().isoformat())
|
t = format(when.astimezone().isoformat())
|
||||||
print("[stat.viewed] ⎩ next update: %s" % (
|
print(
|
||||||
t.split("T")[0] + " " + t.split("T")[1].split(".")[0]
|
"[stat.viewed] ⎩ next update: %s"
|
||||||
))
|
% (t.split("T")[0] + " " + t.split("T")[1].split(".")[0])
|
||||||
|
)
|
||||||
await asyncio.sleep(self.period)
|
await asyncio.sleep(self.period)
|
||||||
else:
|
else:
|
||||||
await asyncio.sleep(10)
|
await asyncio.sleep(10)
|
||||||
|
|
|
@ -3,8 +3,9 @@ from os import environ
|
||||||
PORT = 8080
|
PORT = 8080
|
||||||
|
|
||||||
DB_URL = (
|
DB_URL = (
|
||||||
environ.get("DATABASE_URL") or environ.get("DB_URL") or
|
environ.get("DATABASE_URL")
|
||||||
"postgresql://postgres@localhost:5432/discoursio"
|
or environ.get("DB_URL")
|
||||||
|
or "postgresql://postgres@localhost:5432/discoursio"
|
||||||
)
|
)
|
||||||
JWT_ALGORITHM = "HS256"
|
JWT_ALGORITHM = "HS256"
|
||||||
JWT_SECRET_KEY = environ.get("JWT_SECRET_KEY") or "8f1bd7696ffb482d8486dfbc6e7d16dd-secret-key"
|
JWT_SECRET_KEY = environ.get("JWT_SECRET_KEY") or "8f1bd7696ffb482d8486dfbc6e7d16dd-secret-key"
|
||||||
|
@ -30,4 +31,4 @@ SENTRY_DSN = environ.get("SENTRY_DSN")
|
||||||
SESSION_SECRET_KEY = environ.get("SESSION_SECRET_KEY") or "!secret"
|
SESSION_SECRET_KEY = environ.get("SESSION_SECRET_KEY") or "!secret"
|
||||||
|
|
||||||
# for local development
|
# for local development
|
||||||
DEV_SERVER_PID_FILE_NAME = 'dev-server.pid'
|
DEV_SERVER_PID_FILE_NAME = "dev-server.pid"
|
||||||
|
|
35
setup.cfg
35
setup.cfg
|
@ -1,23 +1,13 @@
|
||||||
[isort]
|
[isort]
|
||||||
# https://github.com/PyCQA/isort
|
# https://github.com/PyCQA/isort
|
||||||
line_length = 120
|
profile = black
|
||||||
multi_line_output = 3
|
|
||||||
include_trailing_comma = true
|
|
||||||
force_grid_wrap = 0
|
|
||||||
use_parentheses = true
|
|
||||||
force_alphabetical_sort = false
|
|
||||||
|
|
||||||
[tool:brunette]
|
|
||||||
# https://github.com/odwyersoftware/brunette
|
|
||||||
line-length = 120
|
|
||||||
single-quotes = false
|
|
||||||
|
|
||||||
[flake8]
|
[flake8]
|
||||||
# https://github.com/PyCQA/flake8
|
# https://github.com/PyCQA/flake8
|
||||||
exclude = .git,__pycache__,.mypy_cache,.vercel
|
exclude = .git,.mypy_cache,schema_types.py
|
||||||
max-line-length = 120
|
max-line-length = 100
|
||||||
max-complexity = 15
|
max-complexity = 10
|
||||||
select = B,C,E,F,W,T4,B9
|
# select = B,C,E,F,W,T4,B9
|
||||||
# E203: Whitespace before ':'
|
# E203: Whitespace before ':'
|
||||||
# E266: Too many leading '#' for block comment
|
# E266: Too many leading '#' for block comment
|
||||||
# E501: Line too long (82 > 79 characters)
|
# E501: Line too long (82 > 79 characters)
|
||||||
|
@ -25,15 +15,12 @@ select = B,C,E,F,W,T4,B9
|
||||||
# W503: Line break occurred before a binary operator
|
# W503: Line break occurred before a binary operator
|
||||||
# F403: 'from module import *' used; unable to detect undefined names
|
# F403: 'from module import *' used; unable to detect undefined names
|
||||||
# C901: Function is too complex
|
# C901: Function is too complex
|
||||||
ignore = E203,E266,E501,E722,W503,F403,C901
|
# ignore = E203,E266,E501,E722,W503,F403,C901
|
||||||
|
extend-ignore = E203
|
||||||
|
|
||||||
[mypy]
|
[mypy]
|
||||||
# https://github.com/python/mypy
|
# https://github.com/python/mypy
|
||||||
ignore_missing_imports = true
|
exclude = schema_types.py
|
||||||
warn_return_any = false
|
explicit_package_bases = true
|
||||||
warn_unused_configs = true
|
check_untyped_defs = true
|
||||||
disallow_untyped_calls = true
|
plugins = sqlmypy
|
||||||
disallow_untyped_defs = true
|
|
||||||
disallow_incomplete_defs = true
|
|
||||||
[mypy-api.*]
|
|
||||||
ignore_errors = true
|
|
||||||
|
|
|
@ -1,4 +1,5 @@
|
||||||
from typing import Optional, Text
|
from typing import Optional, Text
|
||||||
|
|
||||||
from pydantic import BaseModel
|
from pydantic import BaseModel
|
||||||
|
|
||||||
|
|
||||||
|
|
|
@ -1,4 +1,5 @@
|
||||||
from typing import Optional, Text, List
|
from typing import List, Optional, Text
|
||||||
|
|
||||||
from pydantic import BaseModel
|
from pydantic import BaseModel
|
||||||
|
|
||||||
|
|
||||||
|
@ -20,6 +21,7 @@ class Member(BaseModel):
|
||||||
|
|
||||||
|
|
||||||
class Chat(BaseModel):
|
class Chat(BaseModel):
|
||||||
|
id: int
|
||||||
createdAt: int
|
createdAt: int
|
||||||
createdBy: int
|
createdBy: int
|
||||||
users: List[int]
|
users: List[int]
|
||||||
|
|
Loading…
Reference in New Issue
Block a user