2018-09-06 18:35:02 +00:00
|
|
|
import uuid
|
2018-09-13 15:18:23 +00:00
|
|
|
import logging
|
2019-06-20 08:27:07 +00:00
|
|
|
import urllib.parse
|
2018-09-06 18:35:02 +00:00
|
|
|
|
2019-01-11 10:04:11 +00:00
|
|
|
from django.core.cache import cache
|
|
|
|
from django.conf import settings
|
2018-09-13 15:18:23 +00:00
|
|
|
from django.db import transaction, IntegrityError
|
2018-09-22 12:29:30 +00:00
|
|
|
from django.db.models import Q
|
2018-09-13 15:18:23 +00:00
|
|
|
|
|
|
|
from funkwhale_api.common import channels
|
2018-09-06 18:35:02 +00:00
|
|
|
from funkwhale_api.common import utils as funkwhale_utils
|
|
|
|
|
2019-03-05 14:15:37 +00:00
|
|
|
from . import contexts
|
|
|
|
|
2019-01-30 10:54:43 +00:00
|
|
|
recursive_getattr = funkwhale_utils.recursive_getattr
|
|
|
|
|
2018-09-22 12:29:30 +00:00
|
|
|
|
2018-09-13 15:18:23 +00:00
|
|
|
logger = logging.getLogger(__name__)
|
2019-03-05 14:15:37 +00:00
|
|
|
PUBLIC_ADDRESS = contexts.AS.Public
|
2018-09-06 18:35:02 +00:00
|
|
|
|
2018-04-01 20:11:46 +00:00
|
|
|
ACTIVITY_TYPES = [
|
2018-06-09 13:36:16 +00:00
|
|
|
"Accept",
|
|
|
|
"Add",
|
|
|
|
"Announce",
|
|
|
|
"Arrive",
|
|
|
|
"Block",
|
|
|
|
"Create",
|
|
|
|
"Delete",
|
|
|
|
"Dislike",
|
|
|
|
"Flag",
|
|
|
|
"Follow",
|
|
|
|
"Ignore",
|
|
|
|
"Invite",
|
|
|
|
"Join",
|
|
|
|
"Leave",
|
|
|
|
"Like",
|
|
|
|
"Listen",
|
|
|
|
"Move",
|
|
|
|
"Offer",
|
|
|
|
"Question",
|
|
|
|
"Reject",
|
|
|
|
"Read",
|
|
|
|
"Remove",
|
|
|
|
"TentativeReject",
|
|
|
|
"TentativeAccept",
|
|
|
|
"Travel",
|
|
|
|
"Undo",
|
|
|
|
"Update",
|
|
|
|
"View",
|
2018-04-01 20:11:46 +00:00
|
|
|
]
|
|
|
|
|
2018-12-05 16:14:33 +00:00
|
|
|
FUNKWHALE_OBJECT_TYPES = [
|
|
|
|
("Domain", "Domain"),
|
|
|
|
("Artist", "Artist"),
|
|
|
|
("Album", "Album"),
|
|
|
|
("Track", "Track"),
|
|
|
|
("Library", "Library"),
|
|
|
|
]
|
|
|
|
OBJECT_TYPES = (
|
|
|
|
[
|
|
|
|
"Application",
|
|
|
|
"Article",
|
|
|
|
"Audio",
|
|
|
|
"Collection",
|
|
|
|
"Document",
|
|
|
|
"Event",
|
|
|
|
"Group",
|
|
|
|
"Image",
|
|
|
|
"Note",
|
|
|
|
"Object",
|
|
|
|
"OrderedCollection",
|
|
|
|
"Organization",
|
|
|
|
"Page",
|
|
|
|
"Person",
|
|
|
|
"Place",
|
|
|
|
"Profile",
|
|
|
|
"Relationship",
|
|
|
|
"Service",
|
|
|
|
"Tombstone",
|
|
|
|
"Video",
|
|
|
|
]
|
|
|
|
+ ACTIVITY_TYPES
|
|
|
|
+ FUNKWHALE_OBJECT_TYPES
|
|
|
|
)
|
2018-04-01 20:11:46 +00:00
|
|
|
|
2018-04-07 13:34:35 +00:00
|
|
|
|
2018-09-13 15:18:23 +00:00
|
|
|
BROADCAST_TO_USER_ACTIVITIES = ["Follow", "Accept"]
|
2018-09-06 18:35:02 +00:00
|
|
|
|
|
|
|
|
2019-03-05 14:15:37 +00:00
|
|
|
def should_reject(fid, actor_id=None, payload={}):
|
|
|
|
if fid is None and actor_id is None:
|
|
|
|
return False
|
|
|
|
|
2019-01-09 16:52:14 +00:00
|
|
|
from funkwhale_api.moderation import models as moderation_models
|
|
|
|
|
|
|
|
policies = moderation_models.InstancePolicy.objects.active()
|
|
|
|
|
|
|
|
media_types = ["Audio", "Artist", "Album", "Track", "Library", "Image"]
|
|
|
|
relevant_values = [
|
2019-01-30 10:54:43 +00:00
|
|
|
recursive_getattr(payload, "type", permissive=True),
|
|
|
|
recursive_getattr(payload, "object.type", permissive=True),
|
|
|
|
recursive_getattr(payload, "target.type", permissive=True),
|
2019-01-09 16:52:14 +00:00
|
|
|
]
|
|
|
|
# if one of the payload types match our internal media types, then
|
|
|
|
# we apply policies that reject media
|
|
|
|
if set(media_types) & set(relevant_values):
|
|
|
|
policy_type = Q(block_all=True) | Q(reject_media=True)
|
|
|
|
else:
|
|
|
|
policy_type = Q(block_all=True)
|
|
|
|
|
2019-03-05 14:15:37 +00:00
|
|
|
if fid:
|
|
|
|
query = policies.matching_url_query(fid) & policy_type
|
|
|
|
if fid and actor_id:
|
2019-01-09 16:52:14 +00:00
|
|
|
query |= policies.matching_url_query(actor_id) & policy_type
|
2019-03-05 14:15:37 +00:00
|
|
|
elif actor_id:
|
|
|
|
query = policies.matching_url_query(actor_id) & policy_type
|
2019-01-09 16:52:14 +00:00
|
|
|
return policies.filter(query).exists()
|
|
|
|
|
|
|
|
|
2018-09-13 15:18:23 +00:00
|
|
|
@transaction.atomic
|
2018-09-06 18:35:02 +00:00
|
|
|
def receive(activity, on_behalf_of):
|
|
|
|
from . import models
|
|
|
|
from . import serializers
|
|
|
|
from . import tasks
|
2019-03-27 11:53:35 +00:00
|
|
|
from .routes import inbox
|
2019-06-20 08:27:07 +00:00
|
|
|
from funkwhale_api.moderation import mrf
|
2018-09-06 18:35:02 +00:00
|
|
|
|
2019-06-20 08:27:07 +00:00
|
|
|
logger.debug(
|
|
|
|
"[federation] Received activity from %s : %s", on_behalf_of.fid, activity
|
|
|
|
)
|
2018-09-06 18:35:02 +00:00
|
|
|
# we ensure the activity has the bare minimum structure before storing
|
|
|
|
# it in our database
|
|
|
|
serializer = serializers.BaseActivitySerializer(
|
|
|
|
data=activity, context={"actor": on_behalf_of, "local_recipients": True}
|
|
|
|
)
|
|
|
|
serializer.is_valid(raise_exception=True)
|
2019-03-27 11:53:35 +00:00
|
|
|
|
2019-06-20 08:27:07 +00:00
|
|
|
payload, updated = mrf.inbox.apply(activity, sender_id=on_behalf_of.fid)
|
|
|
|
if not payload:
|
2019-01-09 16:52:14 +00:00
|
|
|
logger.info(
|
2019-06-20 08:27:07 +00:00
|
|
|
"[federation] Discarding activity due to mrf %s",
|
2019-01-09 16:52:14 +00:00
|
|
|
serializer.validated_data.get("id"),
|
|
|
|
)
|
|
|
|
return
|
2019-06-20 08:27:07 +00:00
|
|
|
|
|
|
|
if not inbox.get_matching_handlers(payload):
|
|
|
|
# discard unhandlable activity
|
|
|
|
logger.debug(
|
|
|
|
"[federation] No matching route found for activity, discarding: %s", payload
|
|
|
|
)
|
|
|
|
return
|
|
|
|
|
2018-09-13 15:18:23 +00:00
|
|
|
try:
|
2019-06-20 08:27:07 +00:00
|
|
|
copy = serializer.save(payload=payload, type=payload["type"])
|
2018-09-13 15:18:23 +00:00
|
|
|
except IntegrityError:
|
|
|
|
logger.warning(
|
2019-06-20 08:27:07 +00:00
|
|
|
"[federation] Discarding already delivered activity %s",
|
2018-09-13 15:18:23 +00:00
|
|
|
serializer.validated_data.get("id"),
|
|
|
|
)
|
|
|
|
return
|
2018-09-22 12:29:30 +00:00
|
|
|
|
|
|
|
local_to_recipients = get_actors_from_audience(activity.get("to", []))
|
|
|
|
local_to_recipients = local_to_recipients.exclude(user=None)
|
|
|
|
|
|
|
|
local_cc_recipients = get_actors_from_audience(activity.get("cc", []))
|
|
|
|
local_cc_recipients = local_cc_recipients.exclude(user=None)
|
|
|
|
|
|
|
|
inbox_items = []
|
|
|
|
for recipients, type in [(local_to_recipients, "to"), (local_cc_recipients, "cc")]:
|
|
|
|
|
|
|
|
for r in recipients.values_list("pk", flat=True):
|
|
|
|
inbox_items.append(models.InboxItem(actor_id=r, type=type, activity=copy))
|
|
|
|
|
|
|
|
models.InboxItem.objects.bulk_create(inbox_items)
|
|
|
|
|
2018-09-06 18:35:02 +00:00
|
|
|
# at this point, we have the activity in database. Even if we crash, it's
|
|
|
|
# okay, as we can retry later
|
2018-09-13 15:18:23 +00:00
|
|
|
funkwhale_utils.on_commit(tasks.dispatch_inbox.delay, activity_id=copy.pk)
|
2018-09-06 18:35:02 +00:00
|
|
|
return copy
|
|
|
|
|
|
|
|
|
|
|
|
class Router:
|
|
|
|
def __init__(self):
|
|
|
|
self.routes = []
|
|
|
|
|
|
|
|
def connect(self, route, handler):
|
|
|
|
self.routes.append((route, handler))
|
|
|
|
|
|
|
|
def register(self, route):
|
|
|
|
def decorator(handler):
|
|
|
|
self.connect(route, handler)
|
|
|
|
return handler
|
|
|
|
|
|
|
|
return decorator
|
|
|
|
|
|
|
|
|
|
|
|
class InboxRouter(Router):
|
2019-03-06 14:33:39 +00:00
|
|
|
def get_matching_handlers(self, payload):
|
|
|
|
return [
|
|
|
|
handler for route, handler in self.routes if match_route(route, payload)
|
|
|
|
]
|
|
|
|
|
2018-09-13 15:18:23 +00:00
|
|
|
@transaction.atomic
|
2019-03-06 14:33:39 +00:00
|
|
|
def dispatch(self, payload, context, call_handlers=True):
|
2018-09-06 18:35:02 +00:00
|
|
|
"""
|
|
|
|
Receives an Activity payload and some context and trigger our
|
2019-03-06 14:33:39 +00:00
|
|
|
business logic.
|
|
|
|
|
|
|
|
call_handlers should be False when are delivering a local activity, because
|
|
|
|
we want only want to bind activities to their recipients, not reapply the changes.
|
2018-09-06 18:35:02 +00:00
|
|
|
"""
|
2018-09-13 15:18:23 +00:00
|
|
|
from . import api_serializers
|
|
|
|
from . import models
|
|
|
|
|
2019-03-06 14:33:39 +00:00
|
|
|
handlers = self.get_matching_handlers(payload)
|
|
|
|
for handler in handlers:
|
|
|
|
if call_handlers:
|
2018-09-13 15:18:23 +00:00
|
|
|
r = handler(payload, context=context)
|
2019-03-06 14:33:39 +00:00
|
|
|
else:
|
|
|
|
r = None
|
|
|
|
activity_obj = context.get("activity")
|
|
|
|
if activity_obj and r:
|
|
|
|
# handler returned additional data we can use
|
|
|
|
# to update the activity target
|
|
|
|
for key, value in r.items():
|
|
|
|
setattr(activity_obj, key, value)
|
|
|
|
|
|
|
|
update_fields = []
|
|
|
|
for k in r.keys():
|
|
|
|
if k in ["object", "target", "related_object"]:
|
|
|
|
update_fields += [
|
|
|
|
"{}_id".format(k),
|
|
|
|
"{}_content_type".format(k),
|
|
|
|
]
|
|
|
|
else:
|
|
|
|
update_fields.append(k)
|
|
|
|
activity_obj.save(update_fields=update_fields)
|
|
|
|
|
|
|
|
if payload["type"] not in BROADCAST_TO_USER_ACTIVITIES:
|
|
|
|
return
|
2018-09-22 12:29:30 +00:00
|
|
|
|
2019-03-06 14:33:39 +00:00
|
|
|
inbox_items = context.get("inbox_items", models.InboxItem.objects.none())
|
|
|
|
inbox_items = (
|
|
|
|
inbox_items.select_related()
|
|
|
|
.select_related("actor__user")
|
|
|
|
.prefetch_related(
|
|
|
|
"activity__object", "activity__target", "activity__related_object"
|
|
|
|
)
|
|
|
|
)
|
|
|
|
|
|
|
|
for ii in inbox_items:
|
|
|
|
user = ii.actor.get_user()
|
|
|
|
if not user:
|
|
|
|
continue
|
|
|
|
group = "user.{}.inbox".format(user.pk)
|
|
|
|
channels.group_send(
|
|
|
|
group,
|
|
|
|
{
|
|
|
|
"type": "event.send",
|
|
|
|
"text": "",
|
|
|
|
"data": {
|
|
|
|
"type": "inbox.item_added",
|
|
|
|
"item": api_serializers.InboxItemSerializer(ii).data,
|
2018-09-13 15:18:23 +00:00
|
|
|
},
|
2019-03-06 14:33:39 +00:00
|
|
|
},
|
|
|
|
)
|
|
|
|
return
|
2018-09-06 18:35:02 +00:00
|
|
|
|
|
|
|
|
2019-01-11 10:04:11 +00:00
|
|
|
ACTOR_KEY_ROTATION_LOCK_CACHE_KEY = "federation:actor-key-rotation-lock:{}"
|
|
|
|
|
|
|
|
|
|
|
|
def should_rotate_actor_key(actor_id):
|
|
|
|
lock = cache.get(ACTOR_KEY_ROTATION_LOCK_CACHE_KEY.format(actor_id))
|
|
|
|
return lock is None
|
|
|
|
|
|
|
|
|
|
|
|
def schedule_key_rotation(actor_id, delay):
|
|
|
|
from . import tasks
|
|
|
|
|
|
|
|
cache.set(ACTOR_KEY_ROTATION_LOCK_CACHE_KEY.format(actor_id), True, timeout=delay)
|
|
|
|
tasks.rotate_actor_key.apply_async(kwargs={"actor_id": actor_id}, countdown=delay)
|
|
|
|
|
|
|
|
|
2018-09-06 18:35:02 +00:00
|
|
|
class OutboxRouter(Router):
|
2018-09-13 15:18:23 +00:00
|
|
|
@transaction.atomic
|
2018-09-06 18:35:02 +00:00
|
|
|
def dispatch(self, routing, context):
|
|
|
|
"""
|
|
|
|
Receives a routing payload and some business objects in the context
|
|
|
|
and may yield data that should be persisted in the Activity model
|
|
|
|
for further delivery.
|
|
|
|
"""
|
2019-06-20 08:27:07 +00:00
|
|
|
from funkwhale_api.common import preferences
|
2018-09-06 18:35:02 +00:00
|
|
|
from . import models
|
|
|
|
from . import tasks
|
|
|
|
|
2019-06-20 08:27:07 +00:00
|
|
|
allow_list_enabled = preferences.get("moderation__allow_list_enabled")
|
|
|
|
allowed_domains = None
|
|
|
|
if allow_list_enabled:
|
|
|
|
allowed_domains = set(
|
|
|
|
models.Domain.objects.filter(allowed=True).values_list(
|
|
|
|
"name", flat=True
|
|
|
|
)
|
|
|
|
)
|
|
|
|
|
2018-09-06 18:35:02 +00:00
|
|
|
for route, handler in self.routes:
|
2018-09-22 12:29:30 +00:00
|
|
|
if not match_route(route, routing):
|
|
|
|
continue
|
|
|
|
|
|
|
|
activities_data = []
|
|
|
|
for e in handler(context):
|
|
|
|
# a route can yield zero, one or more activity payloads
|
|
|
|
if e:
|
|
|
|
activities_data.append(e)
|
2019-01-11 10:04:11 +00:00
|
|
|
deletions = [
|
|
|
|
a["actor"].id
|
|
|
|
for a in activities_data
|
|
|
|
if a["payload"]["type"] == "Delete"
|
|
|
|
]
|
|
|
|
for actor_id in deletions:
|
|
|
|
# we way need to triggers a blind key rotation
|
|
|
|
if should_rotate_actor_key(actor_id):
|
|
|
|
schedule_key_rotation(actor_id, settings.ACTOR_KEY_ROTATION_DELAY)
|
2018-09-22 12:29:30 +00:00
|
|
|
inbox_items_by_activity_uuid = {}
|
|
|
|
deliveries_by_activity_uuid = {}
|
|
|
|
prepared_activities = []
|
|
|
|
for activity_data in activities_data:
|
|
|
|
activity_data["payload"]["actor"] = activity_data["actor"].fid
|
|
|
|
to = activity_data["payload"].pop("to", [])
|
|
|
|
cc = activity_data["payload"].pop("cc", [])
|
|
|
|
a = models.Activity(**activity_data)
|
|
|
|
a.uuid = uuid.uuid4()
|
|
|
|
to_inbox_items, to_deliveries, new_to = prepare_deliveries_and_inbox_items(
|
2019-06-20 08:27:07 +00:00
|
|
|
to, "to", allowed_domains=allowed_domains
|
2018-09-22 12:29:30 +00:00
|
|
|
)
|
|
|
|
cc_inbox_items, cc_deliveries, new_cc = prepare_deliveries_and_inbox_items(
|
2019-06-20 08:27:07 +00:00
|
|
|
cc, "cc", allowed_domains=allowed_domains
|
2018-09-22 12:29:30 +00:00
|
|
|
)
|
|
|
|
if not any(
|
|
|
|
[to_inbox_items, to_deliveries, cc_inbox_items, cc_deliveries]
|
|
|
|
):
|
|
|
|
continue
|
|
|
|
deliveries_by_activity_uuid[str(a.uuid)] = to_deliveries + cc_deliveries
|
|
|
|
inbox_items_by_activity_uuid[str(a.uuid)] = (
|
|
|
|
to_inbox_items + cc_inbox_items
|
|
|
|
)
|
|
|
|
if new_to:
|
|
|
|
a.payload["to"] = new_to
|
|
|
|
if new_cc:
|
|
|
|
a.payload["cc"] = new_cc
|
|
|
|
prepared_activities.append(a)
|
|
|
|
|
|
|
|
activities = models.Activity.objects.bulk_create(prepared_activities)
|
|
|
|
|
|
|
|
for activity in activities:
|
|
|
|
if str(activity.uuid) in deliveries_by_activity_uuid:
|
|
|
|
for obj in deliveries_by_activity_uuid[str(a.uuid)]:
|
|
|
|
obj.activity = activity
|
|
|
|
|
|
|
|
if str(activity.uuid) in inbox_items_by_activity_uuid:
|
|
|
|
for obj in inbox_items_by_activity_uuid[str(a.uuid)]:
|
|
|
|
obj.activity = activity
|
|
|
|
|
|
|
|
# create all deliveries and items, in bulk
|
|
|
|
models.Delivery.objects.bulk_create(
|
|
|
|
[
|
|
|
|
obj
|
|
|
|
for collection in deliveries_by_activity_uuid.values()
|
|
|
|
for obj in collection
|
|
|
|
]
|
|
|
|
)
|
|
|
|
models.InboxItem.objects.bulk_create(
|
|
|
|
[
|
|
|
|
obj
|
|
|
|
for collection in inbox_items_by_activity_uuid.values()
|
|
|
|
for obj in collection
|
|
|
|
]
|
|
|
|
)
|
|
|
|
|
|
|
|
for a in activities:
|
|
|
|
funkwhale_utils.on_commit(tasks.dispatch_outbox.delay, activity_id=a.pk)
|
|
|
|
return activities
|
|
|
|
|
|
|
|
|
2018-09-06 18:35:02 +00:00
|
|
|
def match_route(route, payload):
|
|
|
|
for key, value in route.items():
|
2019-04-05 14:22:32 +00:00
|
|
|
payload_value = recursive_getattr(payload, key, permissive=True)
|
2019-09-21 14:20:49 +00:00
|
|
|
if isinstance(value, list):
|
|
|
|
if payload_value not in value:
|
|
|
|
return False
|
|
|
|
elif payload_value != value:
|
2018-09-06 18:35:02 +00:00
|
|
|
return False
|
|
|
|
|
|
|
|
return True
|
|
|
|
|
|
|
|
|
2019-06-20 08:27:07 +00:00
|
|
|
def is_allowed_url(url, allowed_domains):
|
|
|
|
return (
|
|
|
|
allowed_domains is None
|
|
|
|
or urllib.parse.urlparse(url).hostname in allowed_domains
|
|
|
|
)
|
|
|
|
|
|
|
|
|
|
|
|
def prepare_deliveries_and_inbox_items(recipient_list, type, allowed_domains=None):
|
2018-09-22 12:29:30 +00:00
|
|
|
"""
|
|
|
|
Given a list of recipients (
|
|
|
|
either actor instances, public adresses, a dictionnary with a "type" and "target"
|
|
|
|
keys for followers collections)
|
|
|
|
returns a list of deliveries, alist of inbox_items and a list
|
|
|
|
of urls to persist in the activity in place of the initial recipient list.
|
|
|
|
"""
|
2018-09-06 18:35:02 +00:00
|
|
|
from . import models
|
|
|
|
|
2019-06-20 08:27:07 +00:00
|
|
|
if allowed_domains is not None:
|
|
|
|
allowed_domains = set(allowed_domains)
|
|
|
|
allowed_domains.add(settings.FEDERATION_HOSTNAME)
|
2018-09-22 12:29:30 +00:00
|
|
|
local_recipients = set()
|
|
|
|
remote_inbox_urls = set()
|
|
|
|
urls = []
|
2018-09-06 18:35:02 +00:00
|
|
|
for r in recipient_list:
|
2018-09-22 12:29:30 +00:00
|
|
|
if isinstance(r, models.Actor):
|
|
|
|
if r.is_local:
|
|
|
|
local_recipients.add(r)
|
|
|
|
else:
|
|
|
|
remote_inbox_urls.add(r.shared_inbox_url or r.inbox_url)
|
|
|
|
urls.append(r.fid)
|
|
|
|
elif r == PUBLIC_ADDRESS:
|
|
|
|
urls.append(r)
|
|
|
|
elif isinstance(r, dict) and r["type"] == "followers":
|
|
|
|
received_follows = (
|
|
|
|
r["target"]
|
|
|
|
.received_follows.filter(approved=True)
|
|
|
|
.select_related("actor__user")
|
|
|
|
)
|
|
|
|
for follow in received_follows:
|
|
|
|
actor = follow.actor
|
|
|
|
if actor.is_local:
|
|
|
|
local_recipients.add(actor)
|
|
|
|
else:
|
|
|
|
remote_inbox_urls.add(actor.shared_inbox_url or actor.inbox_url)
|
|
|
|
urls.append(r["target"].followers_url)
|
|
|
|
|
2019-04-11 08:17:10 +00:00
|
|
|
elif isinstance(r, dict) and r["type"] == "instances_with_followers":
|
|
|
|
# we want to broadcast the activity to other instances service actors
|
|
|
|
# when we have at least one follower from this instance
|
|
|
|
follows = (
|
|
|
|
models.LibraryFollow.objects.filter(approved=True)
|
|
|
|
.exclude(actor__domain_id=settings.FEDERATION_HOSTNAME)
|
|
|
|
.exclude(actor__domain=None)
|
|
|
|
.union(
|
|
|
|
models.Follow.objects.filter(approved=True)
|
|
|
|
.exclude(actor__domain_id=settings.FEDERATION_HOSTNAME)
|
|
|
|
.exclude(actor__domain=None)
|
|
|
|
)
|
|
|
|
)
|
2019-09-21 14:20:49 +00:00
|
|
|
followed_domains = list(follows.values_list("actor__domain_id", flat=True))
|
2019-04-11 08:17:10 +00:00
|
|
|
actors = models.Actor.objects.filter(
|
2019-09-21 14:20:49 +00:00
|
|
|
managed_domains__name__in=followed_domains
|
2019-04-11 08:17:10 +00:00
|
|
|
)
|
2019-09-21 14:20:49 +00:00
|
|
|
values = actors.values("shared_inbox_url", "inbox_url", "domain_id")
|
|
|
|
handled_domains = set()
|
2019-04-11 08:17:10 +00:00
|
|
|
for v in values:
|
|
|
|
remote_inbox_urls.add(v["shared_inbox_url"] or v["inbox_url"])
|
2019-09-21 14:20:49 +00:00
|
|
|
handled_domains.add(v["domain_id"])
|
|
|
|
|
|
|
|
if len(handled_domains) >= len(followed_domains):
|
|
|
|
continue
|
|
|
|
|
|
|
|
# for all remaining domains (probably non-funkwhale instances, with no
|
|
|
|
# service actors), we also pick the latest known actor per domain and send the message
|
|
|
|
# there instead
|
|
|
|
remaining_domains = models.Domain.objects.exclude(name__in=handled_domains)
|
|
|
|
remaining_domains = remaining_domains.filter(name__in=followed_domains)
|
|
|
|
actors = models.Actor.objects.filter(domain__in=remaining_domains)
|
|
|
|
actors = (
|
|
|
|
actors.order_by("domain_id", "-last_fetch_date")
|
|
|
|
.distinct("domain_id")
|
|
|
|
.values("shared_inbox_url", "inbox_url")
|
|
|
|
)
|
|
|
|
for v in actors:
|
|
|
|
remote_inbox_urls.add(v["shared_inbox_url"] or v["inbox_url"])
|
2019-06-20 08:27:07 +00:00
|
|
|
|
|
|
|
deliveries = [
|
|
|
|
models.Delivery(inbox_url=url)
|
|
|
|
for url in remote_inbox_urls
|
|
|
|
if is_allowed_url(url, allowed_domains)
|
|
|
|
]
|
|
|
|
urls = [url for url in urls if is_allowed_url(url, allowed_domains)]
|
2018-09-22 12:29:30 +00:00
|
|
|
inbox_items = [
|
|
|
|
models.InboxItem(actor=actor, type=type) for actor in local_recipients
|
|
|
|
]
|
|
|
|
|
|
|
|
return inbox_items, deliveries, urls
|
|
|
|
|
|
|
|
|
|
|
|
def join_queries_or(left, right):
|
|
|
|
if left:
|
|
|
|
return left | right
|
|
|
|
else:
|
|
|
|
return right
|
|
|
|
|
|
|
|
|
|
|
|
def get_actors_from_audience(urls):
|
|
|
|
"""
|
|
|
|
Given a list of urls such as [
|
|
|
|
"https://hello.world/@bob/followers",
|
|
|
|
"https://eldritch.cafe/@alice/followers",
|
|
|
|
"https://funkwhale.demo/libraries/uuid/followers",
|
|
|
|
]
|
|
|
|
Returns a queryset of actors that are member of the collections
|
|
|
|
listed in the given urls. The urls may contain urls referring
|
|
|
|
to an actor, an actor followers collection or an library followers
|
|
|
|
collection.
|
|
|
|
|
|
|
|
Urls that don't match anything are simply discarded
|
|
|
|
"""
|
|
|
|
from . import models
|
|
|
|
|
|
|
|
queries = {"followed": None, "actors": []}
|
|
|
|
for url in urls:
|
|
|
|
if url == PUBLIC_ADDRESS:
|
|
|
|
continue
|
|
|
|
queries["actors"].append(url)
|
|
|
|
queries["followed"] = join_queries_or(
|
|
|
|
queries["followed"], Q(target__followers_url=url)
|
|
|
|
)
|
|
|
|
final_query = None
|
|
|
|
if queries["actors"]:
|
|
|
|
final_query = join_queries_or(final_query, Q(fid__in=queries["actors"]))
|
|
|
|
if queries["followed"]:
|
|
|
|
actor_follows = models.Follow.objects.filter(queries["followed"], approved=True)
|
|
|
|
final_query = join_queries_or(
|
|
|
|
final_query, Q(pk__in=actor_follows.values_list("actor", flat=True))
|
|
|
|
)
|
|
|
|
|
|
|
|
library_follows = models.LibraryFollow.objects.filter(
|
|
|
|
queries["followed"], approved=True
|
|
|
|
)
|
|
|
|
final_query = join_queries_or(
|
|
|
|
final_query, Q(pk__in=library_follows.values_list("actor", flat=True))
|
|
|
|
)
|
|
|
|
if not final_query:
|
|
|
|
return models.Actor.objects.none()
|
|
|
|
return models.Actor.objects.filter(final_query)
|