forked from forks/microblog.pub
Big cleanup part 2 (#58)
* Cleanup little-boxes stuff * Force html5lib for parsing OG data * Bugfixes
This commit is contained in:
parent
d38c43ebe8
commit
a21121308f
11 changed files with 402 additions and 125 deletions
24
app.py
24
app.py
|
@ -55,6 +55,7 @@ from core.shared import MY_PERSON
|
||||||
from core.shared import _add_answers_to_question
|
from core.shared import _add_answers_to_question
|
||||||
from core.shared import _build_thread
|
from core.shared import _build_thread
|
||||||
from core.shared import _get_ip
|
from core.shared import _get_ip
|
||||||
|
from core.shared import activity_url
|
||||||
from core.shared import back
|
from core.shared import back
|
||||||
from core.shared import csrf
|
from core.shared import csrf
|
||||||
from core.shared import login_required
|
from core.shared import login_required
|
||||||
|
@ -410,7 +411,7 @@ def note_by_id(note_id):
|
||||||
return redirect(url_for("outbox_activity", item_id=note_id))
|
return redirect(url_for("outbox_activity", item_id=note_id))
|
||||||
|
|
||||||
data = DB.activities.find_one(
|
data = DB.activities.find_one(
|
||||||
{"box": Box.OUTBOX.value, "remote_id": back.activity_url(note_id)}
|
{"box": Box.OUTBOX.value, "remote_id": activity_url(note_id)}
|
||||||
)
|
)
|
||||||
if not data:
|
if not data:
|
||||||
abort(404)
|
abort(404)
|
||||||
|
@ -548,7 +549,7 @@ def outbox_detail(item_id):
|
||||||
doc = DB.activities.find_one(
|
doc = DB.activities.find_one(
|
||||||
{
|
{
|
||||||
"box": Box.OUTBOX.value,
|
"box": Box.OUTBOX.value,
|
||||||
"remote_id": back.activity_url(item_id),
|
"remote_id": activity_url(item_id),
|
||||||
"meta.public": True,
|
"meta.public": True,
|
||||||
}
|
}
|
||||||
)
|
)
|
||||||
|
@ -556,27 +557,22 @@ def outbox_detail(item_id):
|
||||||
abort(404)
|
abort(404)
|
||||||
|
|
||||||
if doc["meta"].get("deleted", False):
|
if doc["meta"].get("deleted", False):
|
||||||
obj = ap.parse_activity(doc["activity"])
|
abort(404)
|
||||||
resp = jsonify(**obj.get_tombstone().to_dict())
|
|
||||||
resp.status_code = 410
|
|
||||||
return resp
|
|
||||||
return jsonify(**activity_from_doc(doc))
|
return jsonify(**activity_from_doc(doc))
|
||||||
|
|
||||||
|
|
||||||
@app.route("/outbox/<item_id>/activity")
|
@app.route("/outbox/<item_id>/activity")
|
||||||
def outbox_activity(item_id):
|
def outbox_activity(item_id):
|
||||||
data = find_one_activity(
|
data = find_one_activity(
|
||||||
{**in_outbox(), **by_remote_id(back.activity_url(item_id)), **is_public()}
|
{**in_outbox(), **by_remote_id(activity_url(item_id)), **is_public()}
|
||||||
)
|
)
|
||||||
if not data:
|
if not data:
|
||||||
abort(404)
|
abort(404)
|
||||||
|
|
||||||
obj = activity_from_doc(data)
|
obj = activity_from_doc(data)
|
||||||
if data["meta"].get("deleted", False):
|
if data["meta"].get("deleted", False):
|
||||||
obj = ap.parse_activity(data["activity"])
|
abort(404)
|
||||||
resp = jsonify(**obj.get_object().get_tombstone().to_dict())
|
|
||||||
resp.status_code = 410
|
|
||||||
return resp
|
|
||||||
|
|
||||||
if obj["type"] != ActivityType.CREATE.value:
|
if obj["type"] != ActivityType.CREATE.value:
|
||||||
abort(404)
|
abort(404)
|
||||||
|
@ -590,7 +586,7 @@ def outbox_activity_replies(item_id):
|
||||||
data = DB.activities.find_one(
|
data = DB.activities.find_one(
|
||||||
{
|
{
|
||||||
"box": Box.OUTBOX.value,
|
"box": Box.OUTBOX.value,
|
||||||
"remote_id": back.activity_url(item_id),
|
"remote_id": activity_url(item_id),
|
||||||
"meta.deleted": False,
|
"meta.deleted": False,
|
||||||
"meta.public": True,
|
"meta.public": True,
|
||||||
}
|
}
|
||||||
|
@ -627,7 +623,7 @@ def outbox_activity_likes(item_id):
|
||||||
data = DB.activities.find_one(
|
data = DB.activities.find_one(
|
||||||
{
|
{
|
||||||
"box": Box.OUTBOX.value,
|
"box": Box.OUTBOX.value,
|
||||||
"remote_id": back.activity_url(item_id),
|
"remote_id": activity_url(item_id),
|
||||||
"meta.deleted": False,
|
"meta.deleted": False,
|
||||||
"meta.public": True,
|
"meta.public": True,
|
||||||
}
|
}
|
||||||
|
@ -666,7 +662,7 @@ def outbox_activity_shares(item_id):
|
||||||
data = DB.activities.find_one(
|
data = DB.activities.find_one(
|
||||||
{
|
{
|
||||||
"box": Box.OUTBOX.value,
|
"box": Box.OUTBOX.value,
|
||||||
"remote_id": back.activity_url(item_id),
|
"remote_id": activity_url(item_id),
|
||||||
"meta.deleted": False,
|
"meta.deleted": False,
|
||||||
}
|
}
|
||||||
)
|
)
|
||||||
|
|
|
@ -68,7 +68,7 @@ def admin_logout() -> _Response:
|
||||||
@noindex
|
@noindex
|
||||||
def admin_login() -> _Response:
|
def admin_login() -> _Response:
|
||||||
if session.get("logged_in") is True:
|
if session.get("logged_in") is True:
|
||||||
return redirect(url_for("admin_notifications"))
|
return redirect(url_for("admin.admin_notifications"))
|
||||||
|
|
||||||
devices = [doc["device"] for doc in DB.u2f.find()]
|
devices = [doc["device"] for doc in DB.u2f.find()]
|
||||||
u2f_enabled = True if devices else False
|
u2f_enabled = True if devices else False
|
||||||
|
@ -80,7 +80,7 @@ def admin_login() -> _Response:
|
||||||
if verify_pass(pwd):
|
if verify_pass(pwd):
|
||||||
session["logged_in"] = True
|
session["logged_in"] = True
|
||||||
return redirect(
|
return redirect(
|
||||||
request.args.get("redirect") or url_for("admin_notifications")
|
request.args.get("redirect") or url_for("admin.admin_notifications")
|
||||||
)
|
)
|
||||||
else:
|
else:
|
||||||
abort(403)
|
abort(403)
|
||||||
|
@ -98,7 +98,7 @@ def admin_login() -> _Response:
|
||||||
|
|
||||||
session["logged_in"] = True
|
session["logged_in"] = True
|
||||||
return redirect(
|
return redirect(
|
||||||
request.args.get("redirect") or url_for("admin_notifications")
|
request.args.get("redirect") or url_for("admin.admin_notifications")
|
||||||
)
|
)
|
||||||
else:
|
else:
|
||||||
abort(401)
|
abort(401)
|
||||||
|
|
|
@ -38,7 +38,7 @@ from core.meta import MetaKey
|
||||||
from core.meta import _meta
|
from core.meta import _meta
|
||||||
from core.shared import MY_PERSON
|
from core.shared import MY_PERSON
|
||||||
from core.shared import _Response
|
from core.shared import _Response
|
||||||
from core.shared import back
|
from core.shared import activity_url
|
||||||
from core.shared import csrf
|
from core.shared import csrf
|
||||||
from core.shared import login_required
|
from core.shared import login_required
|
||||||
from core.shared import post_to_outbox
|
from core.shared import post_to_outbox
|
||||||
|
@ -291,7 +291,7 @@ def api_undo() -> _Response:
|
||||||
doc = DB.activities.find_one(
|
doc = DB.activities.find_one(
|
||||||
{
|
{
|
||||||
"box": Box.OUTBOX.value,
|
"box": Box.OUTBOX.value,
|
||||||
"$or": [{"remote_id": back.activity_url(oid)}, {"remote_id": oid}],
|
"$or": [{"remote_id": activity_url(oid)}, {"remote_id": oid}],
|
||||||
}
|
}
|
||||||
)
|
)
|
||||||
if not doc:
|
if not doc:
|
||||||
|
|
|
@ -15,14 +15,16 @@ from requests.exceptions import HTTPError
|
||||||
|
|
||||||
import config
|
import config
|
||||||
from config import DB
|
from config import DB
|
||||||
from core import activitypub
|
|
||||||
from core import gc
|
from core import gc
|
||||||
from core.activitypub import Box
|
from core.activitypub import Box
|
||||||
|
from core.inbox import process_inbox
|
||||||
from core.meta import MetaKey
|
from core.meta import MetaKey
|
||||||
from core.meta import _meta
|
from core.meta import _meta
|
||||||
from core.notifications import set_inbox_flags
|
from core.notifications import set_inbox_flags
|
||||||
|
from core.outbox import process_outbox
|
||||||
from core.shared import MY_PERSON
|
from core.shared import MY_PERSON
|
||||||
from core.shared import _add_answers_to_question
|
from core.shared import _add_answers_to_question
|
||||||
|
from core.shared import _Response
|
||||||
from core.shared import back
|
from core.shared import back
|
||||||
from core.shared import p
|
from core.shared import p
|
||||||
from core.shared import post_to_outbox
|
from core.shared import post_to_outbox
|
||||||
|
@ -43,7 +45,7 @@ class TaskError(Exception):
|
||||||
|
|
||||||
|
|
||||||
@blueprint.route("/task/update_question", methods=["POST"])
|
@blueprint.route("/task/update_question", methods=["POST"])
|
||||||
def task_update_question():
|
def task_update_question() -> _Response:
|
||||||
"""Sends an Update."""
|
"""Sends an Update."""
|
||||||
task = p.parse(flask.request)
|
task = p.parse(flask.request)
|
||||||
app.logger.info(f"task={task!r}")
|
app.logger.info(f"task={task!r}")
|
||||||
|
@ -84,7 +86,7 @@ def task_update_question():
|
||||||
|
|
||||||
|
|
||||||
@blueprint.route("/task/fetch_og_meta", methods=["POST"])
|
@blueprint.route("/task/fetch_og_meta", methods=["POST"])
|
||||||
def task_fetch_og_meta():
|
def task_fetch_og_meta() -> _Response:
|
||||||
task = p.parse(flask.request)
|
task = p.parse(flask.request)
|
||||||
app.logger.info(f"task={task!r}")
|
app.logger.info(f"task={task!r}")
|
||||||
iri = task.payload
|
iri = task.payload
|
||||||
|
@ -123,7 +125,7 @@ def task_fetch_og_meta():
|
||||||
|
|
||||||
|
|
||||||
@blueprint.route("/task/cache_object", methods=["POST"])
|
@blueprint.route("/task/cache_object", methods=["POST"])
|
||||||
def task_cache_object():
|
def task_cache_object() -> _Response:
|
||||||
task = p.parse(flask.request)
|
task = p.parse(flask.request)
|
||||||
app.logger.info(f"task={task!r}")
|
app.logger.info(f"task={task!r}")
|
||||||
iri = task.payload
|
iri = task.payload
|
||||||
|
@ -136,7 +138,8 @@ def task_cache_object():
|
||||||
{
|
{
|
||||||
"$set": {
|
"$set": {
|
||||||
"meta.object": obj.to_dict(embed=True),
|
"meta.object": obj.to_dict(embed=True),
|
||||||
"meta.object_actor": activitypub._actor_to_meta(obj.get_actor()),
|
# FIXME(tsileo): set object actor only if different from actor?
|
||||||
|
"meta.object_actor": obj.get_actor().to_dict(embed=True),
|
||||||
}
|
}
|
||||||
},
|
},
|
||||||
)
|
)
|
||||||
|
@ -151,7 +154,7 @@ def task_cache_object():
|
||||||
|
|
||||||
|
|
||||||
@blueprint.route("/task/finish_post_to_outbox", methods=["POST"]) # noqa:C901
|
@blueprint.route("/task/finish_post_to_outbox", methods=["POST"]) # noqa:C901
|
||||||
def task_finish_post_to_outbox():
|
def task_finish_post_to_outbox() -> _Response:
|
||||||
task = p.parse(flask.request)
|
task = p.parse(flask.request)
|
||||||
app.logger.info(f"task={task!r}")
|
app.logger.info(f"task={task!r}")
|
||||||
iri = task.payload
|
iri = task.payload
|
||||||
|
@ -161,24 +164,7 @@ def task_finish_post_to_outbox():
|
||||||
|
|
||||||
recipients = activity.recipients()
|
recipients = activity.recipients()
|
||||||
|
|
||||||
if activity.has_type(ap.ActivityType.DELETE):
|
process_outbox(activity, {})
|
||||||
back.outbox_delete(MY_PERSON, activity)
|
|
||||||
elif activity.has_type(ap.ActivityType.UPDATE):
|
|
||||||
back.outbox_update(MY_PERSON, activity)
|
|
||||||
elif activity.has_type(ap.ActivityType.CREATE):
|
|
||||||
back.outbox_create(MY_PERSON, activity)
|
|
||||||
elif activity.has_type(ap.ActivityType.ANNOUNCE):
|
|
||||||
back.outbox_announce(MY_PERSON, activity)
|
|
||||||
elif activity.has_type(ap.ActivityType.LIKE):
|
|
||||||
back.outbox_like(MY_PERSON, activity)
|
|
||||||
elif activity.has_type(ap.ActivityType.UNDO):
|
|
||||||
obj = activity.get_object()
|
|
||||||
if obj.has_type(ap.ActivityType.LIKE):
|
|
||||||
back.outbox_undo_like(MY_PERSON, obj)
|
|
||||||
elif obj.has_type(ap.ActivityType.ANNOUNCE):
|
|
||||||
back.outbox_undo_announce(MY_PERSON, obj)
|
|
||||||
elif obj.has_type(ap.ActivityType.FOLLOW):
|
|
||||||
back.undo_new_following(MY_PERSON, obj)
|
|
||||||
|
|
||||||
app.logger.info(f"recipients={recipients}")
|
app.logger.info(f"recipients={recipients}")
|
||||||
activity = ap.clean_activity(activity.to_dict())
|
activity = ap.clean_activity(activity.to_dict())
|
||||||
|
@ -197,7 +183,7 @@ def task_finish_post_to_outbox():
|
||||||
|
|
||||||
|
|
||||||
@blueprint.route("/task/finish_post_to_inbox", methods=["POST"]) # noqa: C901
|
@blueprint.route("/task/finish_post_to_inbox", methods=["POST"]) # noqa: C901
|
||||||
def task_finish_post_to_inbox():
|
def task_finish_post_to_inbox() -> _Response:
|
||||||
task = p.parse(flask.request)
|
task = p.parse(flask.request)
|
||||||
app.logger.info(f"task={task!r}")
|
app.logger.info(f"task={task!r}")
|
||||||
iri = task.payload
|
iri = task.payload
|
||||||
|
@ -205,39 +191,8 @@ def task_finish_post_to_inbox():
|
||||||
activity = ap.fetch_remote_activity(iri)
|
activity = ap.fetch_remote_activity(iri)
|
||||||
app.logger.info(f"activity={activity!r}")
|
app.logger.info(f"activity={activity!r}")
|
||||||
|
|
||||||
if activity.has_type(ap.ActivityType.DELETE):
|
process_inbox(activity, {})
|
||||||
back.inbox_delete(MY_PERSON, activity)
|
|
||||||
elif activity.has_type(ap.ActivityType.UPDATE):
|
|
||||||
back.inbox_update(MY_PERSON, activity)
|
|
||||||
elif activity.has_type(ap.ActivityType.CREATE):
|
|
||||||
back.inbox_create(MY_PERSON, activity)
|
|
||||||
elif activity.has_type(ap.ActivityType.ANNOUNCE):
|
|
||||||
back.inbox_announce(MY_PERSON, activity)
|
|
||||||
elif activity.has_type(ap.ActivityType.LIKE):
|
|
||||||
back.inbox_like(MY_PERSON, activity)
|
|
||||||
elif activity.has_type(ap.ActivityType.FOLLOW):
|
|
||||||
# Reply to a Follow with an Accept
|
|
||||||
actor_id = activity.get_actor().id
|
|
||||||
accept = ap.Accept(
|
|
||||||
actor=config.ID,
|
|
||||||
object={
|
|
||||||
"type": "Follow",
|
|
||||||
"id": activity.id,
|
|
||||||
"object": activity.get_object_id(),
|
|
||||||
"actor": actor_id,
|
|
||||||
},
|
|
||||||
to=[actor_id],
|
|
||||||
published=now(),
|
|
||||||
)
|
|
||||||
post_to_outbox(accept)
|
|
||||||
elif activity.has_type(ap.ActivityType.UNDO):
|
|
||||||
obj = activity.get_object()
|
|
||||||
if obj.has_type(ap.ActivityType.LIKE):
|
|
||||||
back.inbox_undo_like(MY_PERSON, obj)
|
|
||||||
elif obj.has_type(ap.ActivityType.ANNOUNCE):
|
|
||||||
back.inbox_undo_announce(MY_PERSON, obj)
|
|
||||||
elif obj.has_type(ap.ActivityType.FOLLOW):
|
|
||||||
back.undo_new_follower(MY_PERSON, obj)
|
|
||||||
except (ActivityGoneError, ActivityNotFoundError, NotAnActivityError):
|
except (ActivityGoneError, ActivityNotFoundError, NotAnActivityError):
|
||||||
app.logger.exception(f"no retry")
|
app.logger.exception(f"no retry")
|
||||||
except Exception as err:
|
except Exception as err:
|
||||||
|
@ -248,7 +203,7 @@ def task_finish_post_to_inbox():
|
||||||
|
|
||||||
|
|
||||||
@blueprint.route("/task/cache_attachments", methods=["POST"])
|
@blueprint.route("/task/cache_attachments", methods=["POST"])
|
||||||
def task_cache_attachments():
|
def task_cache_attachments() -> _Response:
|
||||||
task = p.parse(flask.request)
|
task = p.parse(flask.request)
|
||||||
app.logger.info(f"task={task!r}")
|
app.logger.info(f"task={task!r}")
|
||||||
iri = task.payload
|
iri = task.payload
|
||||||
|
@ -278,7 +233,7 @@ def task_cache_attachments():
|
||||||
|
|
||||||
|
|
||||||
@blueprint.route("/task/cache_actor", methods=["POST"])
|
@blueprint.route("/task/cache_actor", methods=["POST"])
|
||||||
def task_cache_actor() -> str:
|
def task_cache_actor() -> _Response:
|
||||||
task = p.parse(flask.request)
|
task = p.parse(flask.request)
|
||||||
app.logger.info(f"task={task!r}")
|
app.logger.info(f"task={task!r}")
|
||||||
iri = task.payload["iri"]
|
iri = task.payload["iri"]
|
||||||
|
@ -329,7 +284,7 @@ def task_cache_actor() -> str:
|
||||||
|
|
||||||
|
|
||||||
@blueprint.route("/task/forward_activity", methods=["POST"])
|
@blueprint.route("/task/forward_activity", methods=["POST"])
|
||||||
def task_forward_activity():
|
def task_forward_activity() -> _Response:
|
||||||
task = p.parse(flask.request)
|
task = p.parse(flask.request)
|
||||||
app.logger.info(f"task={task!r}")
|
app.logger.info(f"task={task!r}")
|
||||||
iri = task.payload
|
iri = task.payload
|
||||||
|
@ -350,7 +305,7 @@ def task_forward_activity():
|
||||||
|
|
||||||
|
|
||||||
@blueprint.route("/task/post_to_remote_inbox", methods=["POST"])
|
@blueprint.route("/task/post_to_remote_inbox", methods=["POST"])
|
||||||
def task_post_to_remote_inbox():
|
def task_post_to_remote_inbox() -> _Response:
|
||||||
"""Post an activity to a remote inbox."""
|
"""Post an activity to a remote inbox."""
|
||||||
task = p.parse(flask.request)
|
task = p.parse(flask.request)
|
||||||
app.logger.info(f"task={task!r}")
|
app.logger.info(f"task={task!r}")
|
||||||
|
@ -394,7 +349,7 @@ def task_post_to_remote_inbox():
|
||||||
|
|
||||||
|
|
||||||
@blueprint.route("/task/fetch_remote_question", methods=["POST"])
|
@blueprint.route("/task/fetch_remote_question", methods=["POST"])
|
||||||
def task_fetch_remote_question():
|
def task_fetch_remote_question() -> _Response:
|
||||||
"""Fetch a remote question for implementation that does not send Update."""
|
"""Fetch a remote question for implementation that does not send Update."""
|
||||||
task = p.parse(flask.request)
|
task = p.parse(flask.request)
|
||||||
app.logger.info(f"task={task!r}")
|
app.logger.info(f"task={task!r}")
|
||||||
|
@ -454,7 +409,7 @@ def task_fetch_remote_question():
|
||||||
|
|
||||||
|
|
||||||
@blueprint.route("/task/cleanup", methods=["POST"])
|
@blueprint.route("/task/cleanup", methods=["POST"])
|
||||||
def task_cleanup():
|
def task_cleanup() -> _Response:
|
||||||
task = p.parse(flask.request)
|
task = p.parse(flask.request)
|
||||||
app.logger.info(f"task={task!r}")
|
app.logger.info(f"task={task!r}")
|
||||||
gc.perform()
|
gc.perform()
|
||||||
|
@ -462,7 +417,7 @@ def task_cleanup():
|
||||||
|
|
||||||
|
|
||||||
@blueprint.route("/task/process_new_activity", methods=["POST"]) # noqa:c901
|
@blueprint.route("/task/process_new_activity", methods=["POST"]) # noqa:c901
|
||||||
def task_process_new_activity():
|
def task_process_new_activity() -> _Response:
|
||||||
"""Process an activity received in the inbox"""
|
"""Process an activity received in the inbox"""
|
||||||
task = p.parse(flask.request)
|
task = p.parse(flask.request)
|
||||||
app.logger.info(f"task={task!r}")
|
app.logger.info(f"task={task!r}")
|
||||||
|
|
|
@ -1,5 +1,4 @@
|
||||||
import hashlib
|
import hashlib
|
||||||
import json
|
|
||||||
import logging
|
import logging
|
||||||
import os
|
import os
|
||||||
from datetime import datetime
|
from datetime import datetime
|
||||||
|
@ -33,6 +32,8 @@ from core.tasks import Tasks
|
||||||
|
|
||||||
logger = logging.getLogger(__name__)
|
logger = logging.getLogger(__name__)
|
||||||
|
|
||||||
|
_NewMeta = Dict[str, Any]
|
||||||
|
|
||||||
|
|
||||||
ACTORS_CACHE = LRUCache(maxsize=256)
|
ACTORS_CACHE = LRUCache(maxsize=256)
|
||||||
MY_PERSON = ap.Person(**ME)
|
MY_PERSON = ap.Person(**ME)
|
||||||
|
@ -98,6 +99,9 @@ def _is_local_reply(create: ap.Create) -> bool:
|
||||||
class MicroblogPubBackend(Backend):
|
class MicroblogPubBackend(Backend):
|
||||||
"""Implements a Little Boxes backend, backed by MongoDB."""
|
"""Implements a Little Boxes backend, backed by MongoDB."""
|
||||||
|
|
||||||
|
def base_url(self) -> str:
|
||||||
|
return BASE_URL
|
||||||
|
|
||||||
def debug_mode(self) -> bool:
|
def debug_mode(self) -> bool:
|
||||||
return strtobool(os.getenv("MICROBLOGPUB_DEBUG", "false"))
|
return strtobool(os.getenv("MICROBLOGPUB_DEBUG", "false"))
|
||||||
|
|
||||||
|
@ -108,29 +112,19 @@ class MicroblogPubBackend(Backend):
|
||||||
def extra_inboxes(self) -> List[str]:
|
def extra_inboxes(self) -> List[str]:
|
||||||
return EXTRA_INBOXES
|
return EXTRA_INBOXES
|
||||||
|
|
||||||
def base_url(self) -> str:
|
|
||||||
"""Base URL config."""
|
|
||||||
return BASE_URL
|
|
||||||
|
|
||||||
def activity_url(self, obj_id):
|
|
||||||
"""URL for activity link."""
|
|
||||||
return f"{BASE_URL}/outbox/{obj_id}"
|
|
||||||
|
|
||||||
def note_url(self, obj_id):
|
|
||||||
"""URL for activity link."""
|
|
||||||
return f"{BASE_URL}/note/{obj_id}"
|
|
||||||
|
|
||||||
def save(self, box: Box, activity: ap.BaseActivity) -> None:
|
def save(self, box: Box, activity: ap.BaseActivity) -> None:
|
||||||
"""Custom helper for saving an activity to the DB."""
|
"""Custom helper for saving an activity to the DB."""
|
||||||
visibility = ap.get_visibility(activity)
|
visibility = ap.get_visibility(activity)
|
||||||
is_public = False
|
is_public = False
|
||||||
if visibility in [ap.Visibility.PUBLIC, ap.Visibility.UNLISTED]:
|
if visibility in [ap.Visibility.PUBLIC, ap.Visibility.UNLISTED]:
|
||||||
is_public = True
|
is_public = True
|
||||||
|
|
||||||
object_id = None
|
object_id = None
|
||||||
try:
|
try:
|
||||||
object_id = activity.get_object_id()
|
object_id = activity.get_object_id()
|
||||||
except Exception: # TODO(tsileo): should be ValueError, but replies trigger a KeyError on object
|
except Exception: # TODO(tsileo): should be ValueError, but replies trigger a KeyError on object
|
||||||
pass
|
pass
|
||||||
|
|
||||||
object_visibility = None
|
object_visibility = None
|
||||||
if activity.has_type(
|
if activity.has_type(
|
||||||
[ap.ActivityType.CREATE, ap.ActivityType.ANNOUNCE, ap.ActivityType.LIKE]
|
[ap.ActivityType.CREATE, ap.ActivityType.ANNOUNCE, ap.ActivityType.LIKE]
|
||||||
|
@ -695,25 +689,6 @@ class MicroblogPubBackend(Backend):
|
||||||
{"$set": {"meta.thread_root_parent": root_reply}},
|
{"$set": {"meta.thread_root_parent": root_reply}},
|
||||||
)
|
)
|
||||||
|
|
||||||
def post_to_outbox(self, activity: ap.BaseActivity) -> None:
|
|
||||||
if activity.has_type(ap.CREATE_TYPES):
|
|
||||||
activity = activity.build_create()
|
|
||||||
|
|
||||||
self.save(Box.OUTBOX, activity)
|
|
||||||
|
|
||||||
# Assign create a random ID
|
|
||||||
obj_id = self.random_object_id()
|
|
||||||
activity.set_id(self.activity_url(obj_id), obj_id)
|
|
||||||
|
|
||||||
recipients = activity.recipients()
|
|
||||||
logger.info(f"recipients={recipients}")
|
|
||||||
activity = ap.clean_activity(activity.to_dict())
|
|
||||||
|
|
||||||
payload = json.dumps(activity)
|
|
||||||
for recp in recipients:
|
|
||||||
logger.debug(f"posting to {recp}")
|
|
||||||
self.post_to_remote_inbox(self.get_actor(), payload, recp)
|
|
||||||
|
|
||||||
|
|
||||||
def gen_feed():
|
def gen_feed():
|
||||||
fg = FeedGenerator()
|
fg = FeedGenerator()
|
||||||
|
|
|
@ -17,3 +17,11 @@ class CollectionName(Enum):
|
||||||
|
|
||||||
def find_one_activity(q: _Q) -> _Doc:
|
def find_one_activity(q: _Q) -> _Doc:
|
||||||
return DB[CollectionName.ACTIVITIES.value].find_one(q)
|
return DB[CollectionName.ACTIVITIES.value].find_one(q)
|
||||||
|
|
||||||
|
|
||||||
|
def update_one_activity(q: _Q, update: _Q) -> None:
|
||||||
|
DB[CollectionName.ACTIVITIES.value].update_one(q, update)
|
||||||
|
|
||||||
|
|
||||||
|
def update_many_activities(q: _Q, update: _Q) -> None:
|
||||||
|
DB[CollectionName.ACTIVITIES.value].update_many(q, update)
|
||||||
|
|
181
core/inbox.py
Normal file
181
core/inbox.py
Normal file
|
@ -0,0 +1,181 @@
|
||||||
|
import logging
|
||||||
|
from functools import singledispatch
|
||||||
|
from typing import Any
|
||||||
|
from typing import Dict
|
||||||
|
|
||||||
|
from little_boxes import activitypub as ap
|
||||||
|
from little_boxes.errors import NotAnActivityError
|
||||||
|
|
||||||
|
import config
|
||||||
|
from core.activitypub import _answer_key
|
||||||
|
from core.db import DB
|
||||||
|
from core.meta import Box
|
||||||
|
from core.shared import MY_PERSON
|
||||||
|
from core.shared import back
|
||||||
|
from core.shared import post_to_outbox
|
||||||
|
from core.tasks import Tasks
|
||||||
|
from utils import now
|
||||||
|
|
||||||
|
_logger = logging.getLogger(__name__)
|
||||||
|
|
||||||
|
_NewMeta = Dict[str, Any]
|
||||||
|
|
||||||
|
|
||||||
|
@singledispatch
|
||||||
|
def process_inbox(activity: ap.BaseActivity, new_meta: _NewMeta) -> None:
|
||||||
|
_logger.warning(f"skipping {activity!r}")
|
||||||
|
return None
|
||||||
|
|
||||||
|
|
||||||
|
@process_inbox.register
|
||||||
|
def _delete_process_inbox(delete: ap.Delete, new_meta: _NewMeta) -> None:
|
||||||
|
_logger.info(f"process_inbox activity={delete!r}")
|
||||||
|
obj_id = delete.get_object_id()
|
||||||
|
_logger.debug("delete object={obj_id}")
|
||||||
|
try:
|
||||||
|
obj = ap.fetch_remote_activity(obj_id)
|
||||||
|
_logger.info(f"inbox_delete handle_replies obj={obj!r}")
|
||||||
|
in_reply_to = obj.get_in_reply_to() if obj.inReplyTo else None
|
||||||
|
if obj.has_type(ap.CREATE_TYPES):
|
||||||
|
in_reply_to = ap._get_id(
|
||||||
|
DB.activities.find_one(
|
||||||
|
{"meta.object_id": obj_id, "type": ap.ActivityType.CREATE.value}
|
||||||
|
)["activity"]["object"].get("inReplyTo")
|
||||||
|
)
|
||||||
|
if in_reply_to:
|
||||||
|
back._handle_replies_delete(MY_PERSON, in_reply_to)
|
||||||
|
except Exception:
|
||||||
|
_logger.exception(f"failed to handle delete replies for {obj_id}")
|
||||||
|
|
||||||
|
DB.activities.update_one(
|
||||||
|
{"meta.object_id": obj_id, "type": "Create"}, {"$set": {"meta.deleted": True}}
|
||||||
|
)
|
||||||
|
|
||||||
|
# Foce undo other related activities
|
||||||
|
DB.activities.update({"meta.object_id": obj_id}, {"$set": {"meta.undo": True}})
|
||||||
|
|
||||||
|
|
||||||
|
@process_inbox.register
|
||||||
|
def _update_process_inbox(update: ap.Update, new_meta: _NewMeta) -> None:
|
||||||
|
_logger.info(f"process_inbox activity={update!r}")
|
||||||
|
obj = update.get_object()
|
||||||
|
if obj.ACTIVITY_TYPE == ap.ActivityType.NOTE:
|
||||||
|
DB.activities.update_one(
|
||||||
|
{"activity.object.id": obj.id}, {"$set": {"activity.object": obj.to_dict()}}
|
||||||
|
)
|
||||||
|
elif obj.has_type(ap.ActivityType.QUESTION):
|
||||||
|
choices = obj._data.get("oneOf", obj.anyOf)
|
||||||
|
total_replies = 0
|
||||||
|
_set = {}
|
||||||
|
for choice in choices:
|
||||||
|
answer_key = _answer_key(choice["name"])
|
||||||
|
cnt = choice["replies"]["totalItems"]
|
||||||
|
total_replies += cnt
|
||||||
|
_set[f"meta.question_answers.{answer_key}"] = cnt
|
||||||
|
|
||||||
|
_set["meta.question_replies"] = total_replies
|
||||||
|
|
||||||
|
DB.activities.update_one(
|
||||||
|
{"box": Box.INBOX.value, "activity.object.id": obj.id}, {"$set": _set}
|
||||||
|
)
|
||||||
|
# Also update the cached copies of the question (like Announce and Like)
|
||||||
|
DB.activities.update_many(
|
||||||
|
{"meta.object.id": obj.id}, {"$set": {"meta.object": obj.to_dict()}}
|
||||||
|
)
|
||||||
|
|
||||||
|
# FIXME(tsileo): handle update actor amd inbox_update_note/inbox_update_actor
|
||||||
|
|
||||||
|
|
||||||
|
@process_inbox.register
|
||||||
|
def _create_process_inbox(create: ap.Create, new_meta: _NewMeta) -> None:
|
||||||
|
_logger.info(f"process_inbox activity={create!r}")
|
||||||
|
# If it's a `Quesiion`, trigger an async task for updating it later (by fetching the remote and updating the
|
||||||
|
# local copy)
|
||||||
|
question = create.get_object()
|
||||||
|
if question.has_type(ap.ActivityType.QUESTION):
|
||||||
|
Tasks.fetch_remote_question(question)
|
||||||
|
|
||||||
|
back._handle_replies(MY_PERSON, create)
|
||||||
|
|
||||||
|
|
||||||
|
@process_inbox.register
|
||||||
|
def _announce_process_inbox(announce: ap.Announce, new_meta: _NewMeta) -> None:
|
||||||
|
_logger.info(f"process_inbox activity={announce!r}")
|
||||||
|
# TODO(tsileo): actually drop it without storing it and better logging, also move the check somewhere else
|
||||||
|
# or remove it?
|
||||||
|
try:
|
||||||
|
obj = announce.get_object()
|
||||||
|
except NotAnActivityError:
|
||||||
|
_logger.exception(
|
||||||
|
f'received an Annouce referencing an OStatus notice ({announce._data["object"]}), dropping the message'
|
||||||
|
)
|
||||||
|
return
|
||||||
|
|
||||||
|
if obj.has_type(ap.ActivityType.QUESTION):
|
||||||
|
Tasks.fetch_remote_question(obj)
|
||||||
|
|
||||||
|
DB.activities.update_one(
|
||||||
|
{"remote_id": announce.id},
|
||||||
|
{
|
||||||
|
"$set": {
|
||||||
|
"meta.object": obj.to_dict(embed=True),
|
||||||
|
"meta.object_actor": obj.get_actor().to_dict(embed=True),
|
||||||
|
}
|
||||||
|
},
|
||||||
|
)
|
||||||
|
DB.activities.update_one(
|
||||||
|
{"activity.object.id": obj.id}, {"$inc": {"meta.count_boost": 1}}
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
|
@process_inbox.register
|
||||||
|
def _like_process_inbox(like: ap.Like, new_meta: _NewMeta) -> None:
|
||||||
|
_logger.info(f"process_inbox activity={like!r}")
|
||||||
|
obj = like.get_object()
|
||||||
|
# Update the meta counter if the object is published by the server
|
||||||
|
DB.activities.update_one(
|
||||||
|
{"box": Box.OUTBOX.value, "activity.object.id": obj.id},
|
||||||
|
{"$inc": {"meta.count_like": 1}},
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
|
@process_inbox.register
|
||||||
|
def _follow_process_inbox(activity: ap.Follow, new_meta: _NewMeta) -> None:
|
||||||
|
_logger.info(f"process_inbox activity={activity!r}")
|
||||||
|
# Reply to a Follow with an Accept
|
||||||
|
actor_id = activity.get_actor().id
|
||||||
|
accept = ap.Accept(
|
||||||
|
actor=config.ID,
|
||||||
|
object={
|
||||||
|
"type": "Follow",
|
||||||
|
"id": activity.id,
|
||||||
|
"object": activity.get_object_id(),
|
||||||
|
"actor": actor_id,
|
||||||
|
},
|
||||||
|
to=[actor_id],
|
||||||
|
published=now(),
|
||||||
|
)
|
||||||
|
post_to_outbox(accept)
|
||||||
|
|
||||||
|
|
||||||
|
@process_inbox.register
|
||||||
|
def _undo_process_inbox(activity: ap.Undo, new_meta: _NewMeta) -> None:
|
||||||
|
_logger.info(f"process_inbox activity={activity!r}")
|
||||||
|
obj = activity.get_object()
|
||||||
|
DB.activities.update_one({"remote_id": obj.id}, {"$set": {"meta.undo": True}})
|
||||||
|
if obj.has_type(ap.ActivityType.LIKE):
|
||||||
|
# Update the meta counter if the object is published by the server
|
||||||
|
DB.activities.update_one(
|
||||||
|
{
|
||||||
|
"box": Box.OUTBOX.value,
|
||||||
|
"meta.object_id": obj.get_object_id(),
|
||||||
|
"type": ap.ActivityType.CREATE.value,
|
||||||
|
},
|
||||||
|
{"$inc": {"meta.count_like": -1}},
|
||||||
|
)
|
||||||
|
elif obj.has_type(ap.ActivityType.ANNOUNCE):
|
||||||
|
announced = obj.get_object()
|
||||||
|
# Update the meta counter if the object is published by the server
|
||||||
|
DB.activities.update_one(
|
||||||
|
{"activity.object.id": announced.id}, {"$inc": {"meta.count_boost": -1}}
|
||||||
|
)
|
148
core/outbox.py
Normal file
148
core/outbox.py
Normal file
|
@ -0,0 +1,148 @@
|
||||||
|
import logging
|
||||||
|
from datetime import datetime
|
||||||
|
from functools import singledispatch
|
||||||
|
from typing import Any
|
||||||
|
from typing import Dict
|
||||||
|
|
||||||
|
from little_boxes import activitypub as ap
|
||||||
|
|
||||||
|
from core.db import DB
|
||||||
|
from core.db import find_one_activity
|
||||||
|
from core.db import update_many_activities
|
||||||
|
from core.shared import MY_PERSON
|
||||||
|
from core.shared import back
|
||||||
|
from core.tasks import Tasks
|
||||||
|
|
||||||
|
_logger = logging.getLogger(__name__)
|
||||||
|
|
||||||
|
_NewMeta = Dict[str, Any]
|
||||||
|
|
||||||
|
|
||||||
|
@singledispatch
|
||||||
|
def process_outbox(activity: ap.BaseActivity, new_meta: _NewMeta) -> None:
|
||||||
|
_logger.warning(f"skipping {activity!r}")
|
||||||
|
return None
|
||||||
|
|
||||||
|
|
||||||
|
@process_outbox.register
|
||||||
|
def _delete_process_outbox(delete: ap.Delete, new_meta: _NewMeta) -> None:
|
||||||
|
_logger.info(f"process_outbox activity={delete!r}")
|
||||||
|
obj_id = delete.get_object_id()
|
||||||
|
|
||||||
|
# Flag everything referencing the deleted object as deleted (except the Delete activity itself)
|
||||||
|
update_many_activities(
|
||||||
|
{"meta.object_id": obj_id, "remote_id": {"$ne": delete.id}},
|
||||||
|
{"$set": {"meta.deleted": True, "meta.undo": True}},
|
||||||
|
)
|
||||||
|
|
||||||
|
# If the deleted activity was in DB, decrease some threads-related counter
|
||||||
|
data = find_one_activity(
|
||||||
|
{"meta.object_id": obj_id, "type": ap.ActivityType.CREATE.value}
|
||||||
|
)
|
||||||
|
_logger.info(f"found local copy of deleted activity: {data}")
|
||||||
|
if data:
|
||||||
|
obj = ap.parse_activity(data["activity"]).get_object()
|
||||||
|
_logger.info(f"obj={obj!r}")
|
||||||
|
in_reply_to = obj.get_in_reply_to()
|
||||||
|
if in_reply_to:
|
||||||
|
DB.activities.update_one(
|
||||||
|
{"activity.object.id": in_reply_to},
|
||||||
|
{"$inc": {"meta.count_reply": -1, "meta.count_direct_reply": -1}},
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
|
@process_outbox.register
|
||||||
|
def _update_process_outbox(update: ap.Update, new_meta: _NewMeta) -> None:
|
||||||
|
_logger.info(f"process_outbox activity={update!r}")
|
||||||
|
|
||||||
|
obj = update._data["object"]
|
||||||
|
|
||||||
|
update_prefix = "activity.object."
|
||||||
|
to_update: Dict[str, Any] = {"$set": dict(), "$unset": dict()}
|
||||||
|
to_update["$set"][f"{update_prefix}updated"] = (
|
||||||
|
datetime.utcnow().replace(microsecond=0).isoformat() + "Z"
|
||||||
|
)
|
||||||
|
for k, v in obj.items():
|
||||||
|
if k in ["id", "type"]:
|
||||||
|
continue
|
||||||
|
if v is None:
|
||||||
|
to_update["$unset"][f"{update_prefix}{k}"] = ""
|
||||||
|
else:
|
||||||
|
to_update["$set"][f"{update_prefix}{k}"] = v
|
||||||
|
|
||||||
|
if len(to_update["$unset"]) == 0:
|
||||||
|
del to_update["$unset"]
|
||||||
|
|
||||||
|
_logger.info(f"updating note from outbox {obj!r} {to_update}")
|
||||||
|
DB.activities.update_one({"activity.object.id": obj["id"]}, to_update)
|
||||||
|
# FIXME(tsileo): should send an Update (but not a partial one, to all the note's recipients
|
||||||
|
# (create a new Update with the result of the update, and send it without saving it?)
|
||||||
|
|
||||||
|
|
||||||
|
@process_outbox.register
|
||||||
|
def _create_process_outbox(create: ap.Create, new_meta: _NewMeta) -> None:
|
||||||
|
_logger.info(f"process_outbox activity={create!r}")
|
||||||
|
back._handle_replies(MY_PERSON, create)
|
||||||
|
|
||||||
|
|
||||||
|
@process_outbox.register
|
||||||
|
def _announce_process_outbox(announce: ap.Announce, new_meta: _NewMeta) -> None:
|
||||||
|
_logger.info(f"process_outbox activity={announce!r}")
|
||||||
|
|
||||||
|
obj = announce.get_object()
|
||||||
|
if obj.has_type(ap.ActivityType.QUESTION):
|
||||||
|
Tasks.fetch_remote_question(obj)
|
||||||
|
|
||||||
|
DB.activities.update_one(
|
||||||
|
{"remote_id": announce.id},
|
||||||
|
{
|
||||||
|
"$set": {
|
||||||
|
"meta.object": obj.to_dict(embed=True),
|
||||||
|
"meta.object_actor": obj.get_actor().to_dict(embed=True),
|
||||||
|
}
|
||||||
|
},
|
||||||
|
)
|
||||||
|
|
||||||
|
DB.activities.update_one(
|
||||||
|
{"activity.object.id": obj.id}, {"$set": {"meta.boosted": announce.id}}
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
|
@process_outbox.register
|
||||||
|
def _like_process_outbox(like: ap.Like, new_meta: _NewMeta) -> None:
|
||||||
|
_logger.info(f"process_outbox activity={like!r}")
|
||||||
|
|
||||||
|
obj = like.get_object()
|
||||||
|
if obj.has_type(ap.ActivityType.QUESTION):
|
||||||
|
Tasks.fetch_remote_question(obj)
|
||||||
|
|
||||||
|
DB.activities.update_one(
|
||||||
|
{"activity.object.id": obj.id},
|
||||||
|
{"$inc": {"meta.count_like": 1}, "$set": {"meta.liked": like.id}},
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
|
@process_outbox.register
|
||||||
|
def _undo_process_outbox(undo: ap.Undo, new_meta: _NewMeta) -> None:
|
||||||
|
_logger.info(f"process_outbox activity={undo!r}")
|
||||||
|
obj = undo.get_object()
|
||||||
|
DB.activities.update_one({"remote_id": obj.id}, {"$set": {"meta.undo": True}})
|
||||||
|
|
||||||
|
# Undo Like
|
||||||
|
if obj.has_type(ap.ActivityType.LIKE):
|
||||||
|
liked = obj.get_object_id()
|
||||||
|
DB.activities.update_one(
|
||||||
|
{"activity.object.id": liked},
|
||||||
|
{"$inc": {"meta.count_like": -1}, "$set": {"meta.liked": False}},
|
||||||
|
)
|
||||||
|
|
||||||
|
elif obj.has_type(ap.ActivityType.ANNOUNCE):
|
||||||
|
announced = obj.get_object_id()
|
||||||
|
DB.activities.update_one(
|
||||||
|
{"activity.object.id": announced}, {"$set": {"meta.boosted": False}}
|
||||||
|
)
|
||||||
|
|
||||||
|
# Undo Follow (undo new following)
|
||||||
|
elif obj.has_type(ap.ActivityType.FOLLOW):
|
||||||
|
pass
|
||||||
|
# do nothing
|
|
@ -1,13 +1,13 @@
|
||||||
|
import binascii
|
||||||
import os
|
import os
|
||||||
from datetime import datetime
|
from datetime import datetime
|
||||||
from datetime import timezone
|
from datetime import timezone
|
||||||
from functools import wraps
|
from functools import wraps
|
||||||
from typing import Any
|
from typing import Any
|
||||||
from typing import Dict
|
from typing import Dict
|
||||||
from typing import Union
|
from urllib.parse import urljoin
|
||||||
|
|
||||||
import flask
|
import flask
|
||||||
import werkzeug
|
|
||||||
from bson.objectid import ObjectId
|
from bson.objectid import ObjectId
|
||||||
from flask import current_app as app
|
from flask import current_app as app
|
||||||
from flask import redirect
|
from flask import redirect
|
||||||
|
@ -19,6 +19,7 @@ from little_boxes import activitypub as ap
|
||||||
from little_boxes.activitypub import format_datetime
|
from little_boxes.activitypub import format_datetime
|
||||||
from poussetaches import PousseTaches
|
from poussetaches import PousseTaches
|
||||||
|
|
||||||
|
from config import BASE_URL
|
||||||
from config import DB
|
from config import DB
|
||||||
from config import ME
|
from config import ME
|
||||||
from core import activitypub
|
from core import activitypub
|
||||||
|
@ -26,7 +27,8 @@ from core.activitypub import _answer_key
|
||||||
from core.meta import Box
|
from core.meta import Box
|
||||||
from core.tasks import Tasks
|
from core.tasks import Tasks
|
||||||
|
|
||||||
_Response = Union[flask.Response, werkzeug.wrappers.Response, str]
|
# _Response = Union[flask.Response, werkzeug.wrappers.Response, str, Any]
|
||||||
|
_Response = Any
|
||||||
|
|
||||||
p = PousseTaches(
|
p = PousseTaches(
|
||||||
os.getenv("MICROBLOGPUB_POUSSETACHES_HOST", "http://localhost:7991"),
|
os.getenv("MICROBLOGPUB_POUSSETACHES_HOST", "http://localhost:7991"),
|
||||||
|
@ -69,7 +71,7 @@ def login_required(f):
|
||||||
@wraps(f)
|
@wraps(f)
|
||||||
def decorated_function(*args, **kwargs):
|
def decorated_function(*args, **kwargs):
|
||||||
if not session.get("logged_in"):
|
if not session.get("logged_in"):
|
||||||
return redirect(url_for("admin_login", next=request.url))
|
return redirect(url_for("admin.admin_login", next=request.url))
|
||||||
return f(*args, **kwargs)
|
return f(*args, **kwargs)
|
||||||
|
|
||||||
return decorated_function
|
return decorated_function
|
||||||
|
@ -92,14 +94,26 @@ def _get_ip():
|
||||||
return ip, geoip
|
return ip, geoip
|
||||||
|
|
||||||
|
|
||||||
|
def activity_url(item_id: str) -> str:
|
||||||
|
return urljoin(BASE_URL, url_for("outbox_detail", item_id=item_id))
|
||||||
|
|
||||||
|
|
||||||
def post_to_outbox(activity: ap.BaseActivity) -> str:
|
def post_to_outbox(activity: ap.BaseActivity) -> str:
|
||||||
if activity.has_type(ap.CREATE_TYPES):
|
if activity.has_type(ap.CREATE_TYPES):
|
||||||
activity = activity.build_create()
|
activity = activity.build_create()
|
||||||
|
|
||||||
# Assign create a random ID
|
# Assign create a random ID
|
||||||
obj_id = back.random_object_id()
|
obj_id = binascii.hexlify(os.urandom(8)).decode("utf-8")
|
||||||
|
uri = activity_url(obj_id)
|
||||||
activity.set_id(back.activity_url(obj_id), obj_id)
|
activity._data["id"] = uri
|
||||||
|
if activity.has_type(ap.ActivityType.CREATE):
|
||||||
|
activity._data["object"]["id"] = urljoin(
|
||||||
|
BASE_URL, url_for("outbox_activity", item_id=obj_id)
|
||||||
|
)
|
||||||
|
activity._data["object"]["url"] = urljoin(
|
||||||
|
BASE_URL, url_for("note_by_id", note_id=obj_id)
|
||||||
|
)
|
||||||
|
activity.reset_object_cache()
|
||||||
|
|
||||||
back.save(Box.OUTBOX, activity)
|
back.save(Box.OUTBOX, activity)
|
||||||
Tasks.cache_actor(activity.id)
|
Tasks.cache_actor(activity.id)
|
||||||
|
|
|
@ -23,7 +23,7 @@ def links_from_note(note):
|
||||||
|
|
||||||
links = set()
|
links = set()
|
||||||
if "content" in note:
|
if "content" in note:
|
||||||
soup = BeautifulSoup(note["content"])
|
soup = BeautifulSoup(note["content"], "html5lib")
|
||||||
for link in soup.find_all("a"):
|
for link in soup.find_all("a"):
|
||||||
h = link.get("href")
|
h = link.get("href")
|
||||||
if h.startswith("http") and h not in tags_href and is_url_valid(h):
|
if h.startswith("http") and h not in tags_href and is_url_valid(h):
|
||||||
|
|
|
@ -17,10 +17,10 @@ from little_boxes.activitypub import _to_list
|
||||||
from little_boxes.errors import ActivityGoneError
|
from little_boxes.errors import ActivityGoneError
|
||||||
from little_boxes.errors import ActivityNotFoundError
|
from little_boxes.errors import ActivityNotFoundError
|
||||||
|
|
||||||
from core.activitypub import _answer_key
|
|
||||||
from config import EMOJI_TPL
|
from config import EMOJI_TPL
|
||||||
from config import ID
|
from config import ID
|
||||||
from config import MEDIA_CACHE
|
from config import MEDIA_CACHE
|
||||||
|
from core.activitypub import _answer_key
|
||||||
from utils import parse_datetime
|
from utils import parse_datetime
|
||||||
from utils.media import Kind
|
from utils.media import Kind
|
||||||
|
|
||||||
|
|
Loading…
Reference in a new issue