mirror of
https://git.sr.ht/~tsileo/microblog.pub
synced 2024-11-15 03:04:28 +00:00
Pleroma relay experiment for getting more replies
This commit is contained in:
parent
3904c10c1a
commit
a9fa483bcc
7 changed files with 131 additions and 17 deletions
1
app.py
1
app.py
|
@ -722,6 +722,7 @@ def inbox():
|
||||||
data = remote_data
|
data = remote_data
|
||||||
activity = ap.parse_activity(data)
|
activity = ap.parse_activity(data)
|
||||||
logger.debug(f"inbox activity={g.request_id}/{activity}/{data}")
|
logger.debug(f"inbox activity={g.request_id}/{activity}/{data}")
|
||||||
|
|
||||||
post_to_inbox(activity)
|
post_to_inbox(activity)
|
||||||
|
|
||||||
return Response(status=201)
|
return Response(status=201)
|
||||||
|
|
|
@ -7,6 +7,7 @@ import flask
|
||||||
import requests
|
import requests
|
||||||
from flask import current_app as app
|
from flask import current_app as app
|
||||||
from little_boxes import activitypub as ap
|
from little_boxes import activitypub as ap
|
||||||
|
from little_boxes.activitypub import _to_list
|
||||||
from little_boxes.errors import ActivityGoneError
|
from little_boxes.errors import ActivityGoneError
|
||||||
from little_boxes.errors import ActivityNotFoundError
|
from little_boxes.errors import ActivityNotFoundError
|
||||||
from little_boxes.errors import NotAnActivityError
|
from little_boxes.errors import NotAnActivityError
|
||||||
|
@ -21,10 +22,13 @@ from core.activitypub import Box
|
||||||
from core.activitypub import _actor_hash
|
from core.activitypub import _actor_hash
|
||||||
from core.activitypub import _add_answers_to_question
|
from core.activitypub import _add_answers_to_question
|
||||||
from core.activitypub import post_to_outbox
|
from core.activitypub import post_to_outbox
|
||||||
|
from core.activitypub import save_reply
|
||||||
from core.activitypub import update_cached_actor
|
from core.activitypub import update_cached_actor
|
||||||
|
from core.db import find_one_activity
|
||||||
from core.db import update_one_activity
|
from core.db import update_one_activity
|
||||||
from core.inbox import process_inbox
|
from core.inbox import process_inbox
|
||||||
from core.meta import MetaKey
|
from core.meta import MetaKey
|
||||||
|
from core.meta import by_object_id
|
||||||
from core.meta import by_remote_id
|
from core.meta import by_remote_id
|
||||||
from core.meta import flag
|
from core.meta import flag
|
||||||
from core.meta import upsert
|
from core.meta import upsert
|
||||||
|
@ -310,9 +314,10 @@ def task_cache_actor() -> _Response:
|
||||||
if not activity.has_type([ap.ActivityType.CREATE, ap.ActivityType.ANNOUNCE]):
|
if not activity.has_type([ap.ActivityType.CREATE, ap.ActivityType.ANNOUNCE]):
|
||||||
return ""
|
return ""
|
||||||
|
|
||||||
if activity.get_object()._data.get(
|
if activity.has_type(ap.ActivityType.CREATE) and (
|
||||||
"attachment", []
|
activity.get_object()._data.get("attachment", [])
|
||||||
) or activity.get_object().has_type(ap.ActivityType.VIDEO):
|
or activity.get_object().has_type(ap.ActivityType.VIDEO)
|
||||||
|
):
|
||||||
Tasks.cache_attachments(iri)
|
Tasks.cache_attachments(iri)
|
||||||
|
|
||||||
except (ActivityGoneError, ActivityNotFoundError):
|
except (ActivityGoneError, ActivityNotFoundError):
|
||||||
|
@ -478,6 +483,79 @@ def task_cleanup() -> _Response:
|
||||||
return ""
|
return ""
|
||||||
|
|
||||||
|
|
||||||
|
def _is_local_reply(activity: ap.BaseActivity) -> bool:
|
||||||
|
for dest in _to_list(activity.to or []):
|
||||||
|
if dest.startswith(config.BASE_URL):
|
||||||
|
return True
|
||||||
|
|
||||||
|
for dest in _to_list(activity.cc or []):
|
||||||
|
if dest.startswith(config.BASE_URL):
|
||||||
|
return True
|
||||||
|
|
||||||
|
return False
|
||||||
|
|
||||||
|
|
||||||
|
@blueprint.route("/task/process_reply", methods=["POST"])
|
||||||
|
def task_process_reply() -> _Response:
|
||||||
|
"""Process `Announce`d posts from Pleroma relays in order to process replies of activities that are in the inbox."""
|
||||||
|
task = p.parse(flask.request)
|
||||||
|
app.logger.info(f"task={task!r}")
|
||||||
|
iri = task.payload
|
||||||
|
try:
|
||||||
|
activity = ap.fetch_remote_activity(iri)
|
||||||
|
app.logger.info(f"checking for reply activity={activity!r}")
|
||||||
|
|
||||||
|
# Some AP server always return Create when requesting an object
|
||||||
|
if activity.has_type(ap.ActivityType.CREATE):
|
||||||
|
activity = activity.get_object()
|
||||||
|
|
||||||
|
in_reply_to = activity.get_in_reply_to()
|
||||||
|
if not in_reply_to:
|
||||||
|
# If it's not reply, we can drop it
|
||||||
|
app.logger.info(f"activity={activity!r} is not a reply, dropping it")
|
||||||
|
return ""
|
||||||
|
|
||||||
|
# new_threads = []
|
||||||
|
root_reply = in_reply_to
|
||||||
|
reply = ap.fetch_remote_activity(root_reply)
|
||||||
|
if reply.has_type(ap.ActivityType.CREATE):
|
||||||
|
reply = reply.get_object()
|
||||||
|
|
||||||
|
while reply is not None:
|
||||||
|
in_reply_to = reply.get_in_reply_to()
|
||||||
|
if not in_reply_to:
|
||||||
|
break
|
||||||
|
root_reply = in_reply_to
|
||||||
|
reply = ap.fetch_remote_activity(root_reply)
|
||||||
|
if reply.has_type(ap.ActivityType.CREATE):
|
||||||
|
reply = reply.get_object()
|
||||||
|
|
||||||
|
app.logger.info(f"root_reply={reply!r} for activity={activity!r}")
|
||||||
|
|
||||||
|
# Ensure the "root reply" is present in the inbox/outbox
|
||||||
|
if not find_one_activity(by_object_id(root_reply)):
|
||||||
|
return ""
|
||||||
|
|
||||||
|
actor = activity.get_actor()
|
||||||
|
|
||||||
|
save_reply(
|
||||||
|
activity,
|
||||||
|
{
|
||||||
|
"meta.thread_root_parent": root_reply,
|
||||||
|
**flag(MetaKey.ACTOR, actor.to_dict(embed=True)),
|
||||||
|
},
|
||||||
|
)
|
||||||
|
# FIXME(tsileo): cache actor here, spawn a task to cache attachment if needed
|
||||||
|
except (ActivityGoneError, ActivityNotFoundError):
|
||||||
|
app.logger.exception(f"dropping activity {iri}, skip processing")
|
||||||
|
return ""
|
||||||
|
except Exception as err:
|
||||||
|
app.logger.exception(f"failed to process new activity {iri}")
|
||||||
|
raise TaskError() from err
|
||||||
|
|
||||||
|
return ""
|
||||||
|
|
||||||
|
|
||||||
@blueprint.route("/task/process_new_activity", methods=["POST"]) # noqa:c901
|
@blueprint.route("/task/process_new_activity", methods=["POST"]) # noqa:c901
|
||||||
def task_process_new_activity() -> _Response:
|
def task_process_new_activity() -> _Response:
|
||||||
"""Process an activity received in the inbox"""
|
"""Process an activity received in the inbox"""
|
||||||
|
|
|
@ -12,7 +12,6 @@ from urllib.parse import urljoin
|
||||||
from urllib.parse import urlparse
|
from urllib.parse import urlparse
|
||||||
|
|
||||||
from bson.objectid import ObjectId
|
from bson.objectid import ObjectId
|
||||||
from cachetools import LRUCache
|
|
||||||
from flask import url_for
|
from flask import url_for
|
||||||
from little_boxes import activitypub as ap
|
from little_boxes import activitypub as ap
|
||||||
from little_boxes import strtobool
|
from little_boxes import strtobool
|
||||||
|
@ -43,7 +42,6 @@ _NewMeta = Dict[str, Any]
|
||||||
|
|
||||||
SIG_AUTH = HTTPSigAuth(KEY)
|
SIG_AUTH = HTTPSigAuth(KEY)
|
||||||
|
|
||||||
ACTORS_CACHE = LRUCache(maxsize=256)
|
|
||||||
MY_PERSON = ap.Person(**ME)
|
MY_PERSON = ap.Person(**ME)
|
||||||
|
|
||||||
|
|
||||||
|
@ -157,6 +155,16 @@ def post_to_inbox(activity: ap.BaseActivity) -> None:
|
||||||
)
|
)
|
||||||
return
|
return
|
||||||
|
|
||||||
|
# If the message is coming from a Pleroma relay, we process it as a possible reply for a stream activity
|
||||||
|
if (
|
||||||
|
actor.has_type(ap.ActivityType.APPLICATION)
|
||||||
|
and actor.id.endswith("/relay")
|
||||||
|
and activity.has_type(ap.ActivityType.ANNOUNCE)
|
||||||
|
and not DB.replies.find_one({"remote_id": activity.id})
|
||||||
|
):
|
||||||
|
Tasks.process_reply(activity.get_object_id())
|
||||||
|
return
|
||||||
|
|
||||||
if DB.activities.find_one({"box": Box.INBOX.value, "remote_id": activity.id}):
|
if DB.activities.find_one({"box": Box.INBOX.value, "remote_id": activity.id}):
|
||||||
# The activity is already in the inbox
|
# The activity is already in the inbox
|
||||||
logger.info(f"received duplicate activity {activity!r}, dropping it")
|
logger.info(f"received duplicate activity {activity!r}, dropping it")
|
||||||
|
@ -170,6 +178,30 @@ def post_to_inbox(activity: ap.BaseActivity) -> None:
|
||||||
Tasks.finish_post_to_inbox(activity.id)
|
Tasks.finish_post_to_inbox(activity.id)
|
||||||
|
|
||||||
|
|
||||||
|
def save_reply(activity: ap.BaseActivity, meta: Dict[str, Any] = {}) -> None:
|
||||||
|
visibility = ap.get_visibility(activity)
|
||||||
|
is_public = False
|
||||||
|
if visibility in [ap.Visibility.PUBLIC, ap.Visibility.UNLISTED]:
|
||||||
|
is_public = True
|
||||||
|
|
||||||
|
DB.replies.insert_one(
|
||||||
|
{
|
||||||
|
"activity": activity.to_dict(),
|
||||||
|
"type": _to_list(activity.type),
|
||||||
|
"remote_id": activity.id,
|
||||||
|
"meta": {
|
||||||
|
"undo": False,
|
||||||
|
"deleted": False,
|
||||||
|
"public": is_public,
|
||||||
|
"server": urlparse(activity.id).netloc,
|
||||||
|
"visibility": visibility.name,
|
||||||
|
"actor_id": activity.get_actor().id,
|
||||||
|
**meta,
|
||||||
|
},
|
||||||
|
}
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
def post_to_outbox(activity: ap.BaseActivity) -> str:
|
def post_to_outbox(activity: ap.BaseActivity) -> str:
|
||||||
if activity.has_type(ap.CREATE_TYPES):
|
if activity.has_type(ap.CREATE_TYPES):
|
||||||
activity = activity.build_create()
|
activity = activity.build_create()
|
||||||
|
|
|
@ -98,4 +98,4 @@ def upsert(data: Dict[MetaKey, Any]) -> _SubQuery:
|
||||||
|
|
||||||
|
|
||||||
def published_after(dt: datetime) -> _SubQuery:
|
def published_after(dt: datetime) -> _SubQuery:
|
||||||
return flag(MetaKey.PUBLISHED, {"gt": ap.format(dt)})
|
return flag(MetaKey.PUBLISHED, {"gt": ap.format_datetime(dt)})
|
||||||
|
|
|
@ -140,16 +140,16 @@ def _announce_set_inbox_flags(activity: ap.Announce, new_meta: _NewMeta) -> None
|
||||||
_set_flag(new_meta, MetaKey.GC_KEEP)
|
_set_flag(new_meta, MetaKey.GC_KEEP)
|
||||||
|
|
||||||
# Dedup boosts (it's annoying to see the same note multipe times on the same page)
|
# Dedup boosts (it's annoying to see the same note multipe times on the same page)
|
||||||
# if not find_one_activity(
|
if not find_one_activity(
|
||||||
# {
|
{
|
||||||
# **in_inbox(),
|
**in_inbox(),
|
||||||
# **by_object_id(obj.id),
|
**by_object_id(obj.id),
|
||||||
# **flag(MetaKey.STREAM, True),
|
**flag(MetaKey.STREAM, True),
|
||||||
# **published_after(datetime.now(timezone.utc) - timedelta(hours=12)),
|
**published_after(datetime.now(timezone.utc) - timedelta(hours=12)),
|
||||||
# }
|
}
|
||||||
# ):
|
):
|
||||||
# Display it in the stream only it not there already (only looking at the last 12 hours)
|
# Display it in the stream only it not there already (only looking at the last 12 hours)
|
||||||
_set_flag(new_meta, MetaKey.STREAM)
|
_set_flag(new_meta, MetaKey.STREAM)
|
||||||
|
|
||||||
return None
|
return None
|
||||||
|
|
||||||
|
|
|
@ -44,6 +44,10 @@ class Tasks:
|
||||||
def fetch_og_meta(iri: str) -> None:
|
def fetch_og_meta(iri: str) -> None:
|
||||||
p.push(iri, "/task/fetch_og_meta")
|
p.push(iri, "/task/fetch_og_meta")
|
||||||
|
|
||||||
|
@staticmethod
|
||||||
|
def process_reply(iri: str) -> None:
|
||||||
|
p.push(iri, "/task/process_reply")
|
||||||
|
|
||||||
@staticmethod
|
@staticmethod
|
||||||
def process_new_activity(iri: str) -> None:
|
def process_new_activity(iri: str) -> None:
|
||||||
p.push(iri, "/task/process_new_activity")
|
p.push(iri, "/task/process_new_activity")
|
||||||
|
|
|
@ -21,6 +21,5 @@ git+https://github.com/erikriver/opengraph.git
|
||||||
git+https://github.com/tsileo/little-boxes.git@litepub
|
git+https://github.com/tsileo/little-boxes.git@litepub
|
||||||
pyyaml
|
pyyaml
|
||||||
pillow
|
pillow
|
||||||
cachetools
|
|
||||||
emoji-unicode
|
emoji-unicode
|
||||||
html5lib
|
html5lib
|
||||||
|
|
Loading…
Reference in a new issue