2024-08-28 05:44:18 +00:00
|
|
|
"""ATProto firehose client. Enqueues receive tasks for events for bridged users."""
|
2024-05-08 22:48:39 +00:00
|
|
|
from collections import namedtuple
|
2024-05-14 05:02:09 +00:00
|
|
|
from datetime import datetime, timedelta
|
2024-09-01 14:46:15 +00:00
|
|
|
from io import BytesIO
|
2024-05-07 23:17:44 +00:00
|
|
|
import itertools
|
|
|
|
|
import logging
|
2024-05-07 21:15:51 +00:00
|
|
|
import os
|
2024-08-30 04:05:33 +00:00
|
|
|
from queue import Queue
|
2024-05-10 04:25:13 +00:00
|
|
|
from threading import Event, Lock, Thread, Timer
|
2024-09-01 14:46:15 +00:00
|
|
|
import threading
|
2024-05-09 03:43:10 +00:00
|
|
|
import time
|
2024-05-07 21:15:51 +00:00
|
|
|
|
2024-05-10 18:31:47 +00:00
|
|
|
from arroba.datastore_storage import AtpRepo
|
2024-07-06 22:26:20 +00:00
|
|
|
from arroba.util import parse_at_uri
|
2024-05-07 21:15:51 +00:00
|
|
|
from carbox import read_car
|
2024-09-01 14:46:15 +00:00
|
|
|
import dag_cbor
|
2024-05-07 21:15:51 +00:00
|
|
|
import dag_json
|
2024-05-09 21:31:57 +00:00
|
|
|
from google.cloud import ndb
|
2024-05-07 21:15:51 +00:00
|
|
|
from granary.bluesky import AT_URI_PATTERN
|
|
|
|
|
from lexrpc.client import Client
|
2024-09-01 14:46:15 +00:00
|
|
|
import libipld
|
2024-05-08 17:39:03 +00:00
|
|
|
from oauth_dropins.webutil import util
|
2024-05-09 21:31:57 +00:00
|
|
|
from oauth_dropins.webutil.appengine_config import ndb_client
|
2024-06-13 00:04:59 +00:00
|
|
|
from oauth_dropins.webutil.appengine_info import DEBUG
|
2024-07-05 04:11:38 +00:00
|
|
|
from oauth_dropins.webutil.util import json_dumps, json_loads
|
2024-05-07 21:15:51 +00:00
|
|
|
|
2024-05-09 04:35:03 +00:00
|
|
|
from atproto import ATProto, Cursor
|
2024-06-05 02:10:01 +00:00
|
|
|
from common import (
|
|
|
|
|
add,
|
2024-08-01 19:07:20 +00:00
|
|
|
cache_policy,
|
2024-06-05 02:10:01 +00:00
|
|
|
create_task,
|
|
|
|
|
global_cache,
|
2024-08-01 19:42:13 +00:00
|
|
|
global_cache_policy,
|
2024-06-05 02:10:01 +00:00
|
|
|
global_cache_timeout_policy,
|
|
|
|
|
report_exception,
|
|
|
|
|
USER_AGENT,
|
|
|
|
|
)
|
2024-05-14 05:02:09 +00:00
|
|
|
from models import Object, reset_protocol_properties
|
|
|
|
|
|
2024-06-13 00:04:59 +00:00
|
|
|
logger = logging.getLogger(__name__)
|
2024-05-07 23:17:44 +00:00
|
|
|
|
2024-05-09 03:43:10 +00:00
|
|
|
RECONNECT_DELAY = timedelta(seconds=30)
|
2024-05-23 05:08:25 +00:00
|
|
|
STORE_CURSOR_FREQ = timedelta(seconds=10)
|
2024-05-09 03:43:10 +00:00
|
|
|
|
2024-05-08 22:48:39 +00:00
|
|
|
# a commit operation. similar to arroba.repo.Write. record is None for deletes.
|
2024-05-09 06:11:27 +00:00
|
|
|
Op = namedtuple('Op', ['action', 'repo', 'path', 'seq', 'record'],
|
2024-05-08 22:48:39 +00:00
|
|
|
# record is optional
|
|
|
|
|
defaults=[None])
|
|
|
|
|
|
2024-09-01 14:46:15 +00:00
|
|
|
# contains bytes, websocket frames
|
2024-08-30 04:05:33 +00:00
|
|
|
#
|
|
|
|
|
# maxsize is important here! if we hit this limit, subscribe will block when it
|
|
|
|
|
# tries to add more commits until handle consumes some. this keeps subscribe
|
|
|
|
|
# from getting too far ahead of handle and using too much memory in this queue.
|
2024-09-01 14:46:15 +00:00
|
|
|
events = Queue(maxsize=200)
|
2024-05-08 17:39:03 +00:00
|
|
|
|
2024-05-19 20:47:13 +00:00
|
|
|
# global so that subscribe can reuse it across calls
|
2024-08-30 20:16:17 +00:00
|
|
|
cursor = None
|
2024-05-19 20:47:13 +00:00
|
|
|
|
|
|
|
|
# global: _load_dids populates them, subscribe and handle use them
|
2024-05-14 05:02:09 +00:00
|
|
|
atproto_dids = set()
|
|
|
|
|
atproto_loaded_at = datetime(1900, 1, 1)
|
|
|
|
|
bridged_dids = set()
|
|
|
|
|
bridged_loaded_at = datetime(1900, 1, 1)
|
2024-05-10 04:25:13 +00:00
|
|
|
dids_initialized = Event()
|
2024-05-09 21:31:57 +00:00
|
|
|
|
2024-05-07 21:15:51 +00:00
|
|
|
|
2024-05-09 21:05:05 +00:00
|
|
|
def load_dids():
|
2024-05-11 02:59:04 +00:00
|
|
|
# run in a separate thread since it needs to make its own NDB
|
|
|
|
|
# context when it runs in the timer thread
|
|
|
|
|
Thread(target=_load_dids).start()
|
|
|
|
|
dids_initialized.wait()
|
2024-05-25 01:32:32 +00:00
|
|
|
dids_initialized.clear()
|
2024-05-09 21:31:57 +00:00
|
|
|
|
|
|
|
|
|
|
|
|
|
def _load_dids():
|
2024-05-14 05:02:09 +00:00
|
|
|
global atproto_dids, atproto_loaded_at, bridged_dids, bridged_loaded_at
|
2024-05-09 03:43:10 +00:00
|
|
|
|
2024-08-01 19:07:20 +00:00
|
|
|
with ndb_client.context(cache_policy=cache_policy, global_cache=global_cache,
|
2024-08-01 19:42:13 +00:00
|
|
|
global_cache_policy=global_cache_policy,
|
2024-06-05 02:10:01 +00:00
|
|
|
global_cache_timeout_policy=global_cache_timeout_policy):
|
2024-05-09 22:38:52 +00:00
|
|
|
if not DEBUG:
|
2024-05-10 04:25:13 +00:00
|
|
|
Timer(STORE_CURSOR_FREQ.total_seconds(), _load_dids).start()
|
2024-05-09 03:43:10 +00:00
|
|
|
|
2024-05-14 05:02:09 +00:00
|
|
|
atproto_query = ATProto.query(ATProto.enabled_protocols != None,
|
2024-05-25 01:32:32 +00:00
|
|
|
ATProto.updated > atproto_loaded_at)
|
2024-07-20 22:19:18 +00:00
|
|
|
atproto_loaded_at = ATProto.query().order(-ATProto.updated).get().updated
|
2024-05-14 05:02:09 +00:00
|
|
|
new_atproto = [key.id() for key in atproto_query.iter(keys_only=True)]
|
|
|
|
|
atproto_dids.update(new_atproto)
|
|
|
|
|
|
2024-07-23 05:24:13 +00:00
|
|
|
bridged_query = AtpRepo.query(AtpRepo.status == None,
|
|
|
|
|
AtpRepo.created > bridged_loaded_at)
|
2024-05-14 05:02:09 +00:00
|
|
|
bridged_loaded_at = AtpRepo.query().order(-AtpRepo.created).get().created
|
|
|
|
|
new_bridged = [key.id() for key in bridged_query.iter(keys_only=True)]
|
|
|
|
|
bridged_dids.update(new_bridged)
|
2024-05-07 21:15:51 +00:00
|
|
|
|
2024-05-10 04:25:13 +00:00
|
|
|
dids_initialized.set()
|
2024-09-01 07:26:37 +00:00
|
|
|
total = len(atproto_dids) + len(bridged_dids)
|
2024-09-01 14:46:15 +00:00
|
|
|
logger.info(f'DIDs: {total} ATProto {len(atproto_dids)} (+{len(new_atproto)}), AtpRepo {len(bridged_dids)} (+{len(new_bridged)}); events {events.qsize()}')
|
2024-05-09 21:05:05 +00:00
|
|
|
|
|
|
|
|
|
2024-05-14 05:02:09 +00:00
|
|
|
def subscriber():
|
|
|
|
|
"""Wrapper around :func:`_subscribe` that catches exceptions and reconnects."""
|
2024-05-09 21:05:05 +00:00
|
|
|
logger.info(f'started thread to subscribe to {os.environ["BGS_HOST"]} firehose')
|
2024-05-25 01:32:32 +00:00
|
|
|
load_dids()
|
2024-05-07 21:15:51 +00:00
|
|
|
|
2024-05-09 03:43:10 +00:00
|
|
|
while True:
|
|
|
|
|
try:
|
2024-06-05 02:10:01 +00:00
|
|
|
with ndb_client.context(
|
2024-08-01 19:07:20 +00:00
|
|
|
cache_policy=cache_policy, global_cache=global_cache,
|
2024-08-01 19:42:13 +00:00
|
|
|
global_cache_policy=global_cache_policy,
|
2024-08-01 19:07:20 +00:00
|
|
|
global_cache_timeout_policy=global_cache_timeout_policy):
|
2024-05-14 05:02:09 +00:00
|
|
|
subscribe()
|
|
|
|
|
|
|
|
|
|
logger.info(f'disconnected! waiting {RECONNECT_DELAY} and then reconnecting')
|
|
|
|
|
time.sleep(RECONNECT_DELAY.total_seconds())
|
|
|
|
|
|
2024-05-10 04:31:41 +00:00
|
|
|
except BaseException:
|
|
|
|
|
report_exception()
|
2024-05-09 03:43:10 +00:00
|
|
|
|
|
|
|
|
|
2024-05-14 05:02:09 +00:00
|
|
|
def subscribe():
|
2024-05-11 02:59:04 +00:00
|
|
|
"""Subscribes to the relay's firehose.
|
|
|
|
|
|
|
|
|
|
Relay hostname comes from the ``BGS_HOST`` environment variable.
|
|
|
|
|
|
|
|
|
|
Args:
|
|
|
|
|
reconnect (bool): whether to always reconnect after we get disconnected
|
|
|
|
|
"""
|
2024-08-30 20:16:17 +00:00
|
|
|
global cursor
|
|
|
|
|
if not cursor:
|
|
|
|
|
cursor = Cursor.get_or_insert(
|
2024-05-19 20:47:13 +00:00
|
|
|
f'{os.environ["BGS_HOST"]} com.atproto.sync.subscribeRepos')
|
2024-08-30 20:16:17 +00:00
|
|
|
# TODO: remove? does this make us skip events? if we remove it, will we
|
|
|
|
|
# infinite loop when we fail on an event?
|
|
|
|
|
if cursor.cursor:
|
|
|
|
|
cursor.cursor += 1
|
2024-05-09 06:11:27 +00:00
|
|
|
|
2024-06-05 02:10:01 +00:00
|
|
|
client = Client(f'https://{os.environ["BGS_HOST"]}',
|
|
|
|
|
headers={'User-Agent': USER_AGENT})
|
2024-05-07 23:58:52 +00:00
|
|
|
|
2024-09-01 14:46:15 +00:00
|
|
|
for frame in client.com.atproto.sync.subscribeRepos(cursor=cursor.cursor):
|
2024-05-14 05:02:09 +00:00
|
|
|
# ops = ' '.join(f'{op.get("action")} {op.get("path")}'
|
|
|
|
|
# for op in payload.get('ops', []))
|
|
|
|
|
# logger.info(f'seeing {payload.get("seq")} {repo} {ops}')
|
|
|
|
|
|
2024-09-01 14:46:15 +00:00
|
|
|
events.put(frame)
|
2024-09-01 07:26:37 +00:00
|
|
|
|
|
|
|
|
|
|
|
|
|
def handler():
|
|
|
|
|
"""Wrapper around :func:`handle` that catches exceptions and restarts."""
|
|
|
|
|
logger.info(f'started handle thread to store objects and enqueue receive tasks')
|
|
|
|
|
|
|
|
|
|
while True:
|
|
|
|
|
try:
|
|
|
|
|
with ndb_client.context(
|
|
|
|
|
cache_policy=cache_policy, global_cache=global_cache,
|
|
|
|
|
global_cache_policy=global_cache_policy,
|
|
|
|
|
global_cache_timeout_policy=global_cache_timeout_policy):
|
|
|
|
|
handle()
|
|
|
|
|
|
|
|
|
|
# if we return cleanly, that means we hit the limit
|
|
|
|
|
break
|
|
|
|
|
|
|
|
|
|
except BaseException:
|
|
|
|
|
report_exception()
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
def handle(limit=None):
|
|
|
|
|
seen = 0
|
|
|
|
|
|
|
|
|
|
def _handle(op):
|
|
|
|
|
type, _ = op.path.strip('/').split('/', maxsplit=1)
|
|
|
|
|
record_json = dag_json.encode(op.record, dialect='atproto')
|
|
|
|
|
if type not in ATProto.SUPPORTED_RECORD_TYPES:
|
2024-09-01 14:46:15 +00:00
|
|
|
logger.info(f'Skipping unsupported type {type}: {record_json}[:1000]')
|
2024-09-01 07:26:37 +00:00
|
|
|
return
|
|
|
|
|
|
|
|
|
|
at_uri = f'at://{op.repo}/{op.path}'
|
|
|
|
|
|
|
|
|
|
# store object, enqueue receive task
|
|
|
|
|
if op.action in ('create', 'update'):
|
|
|
|
|
record_kwarg = {
|
|
|
|
|
'bsky': json_loads(record_json),
|
|
|
|
|
}
|
|
|
|
|
obj_id = at_uri
|
|
|
|
|
elif op.action == 'delete':
|
|
|
|
|
verb = ('delete'
|
|
|
|
|
if type in ('app.bsky.actor.profile', 'app.bsky.feed.post')
|
|
|
|
|
else 'undo')
|
|
|
|
|
obj_id = f'{at_uri}#{verb}'
|
|
|
|
|
record_kwarg = {'our_as1': {
|
|
|
|
|
'objectType': 'activity',
|
|
|
|
|
'verb': verb,
|
|
|
|
|
'id': obj_id,
|
|
|
|
|
'actor': op.repo,
|
|
|
|
|
'object': at_uri,
|
|
|
|
|
}}
|
|
|
|
|
else:
|
|
|
|
|
logger.error(f'Unknown action {action} for {op.repo} {op.path}')
|
|
|
|
|
return
|
|
|
|
|
|
|
|
|
|
try:
|
2024-09-01 14:46:15 +00:00
|
|
|
obj = Object.get_or_create(id=obj_id, authed_as=op.repo, status='new',
|
|
|
|
|
users=[ATProto(id=op.repo).key],
|
|
|
|
|
source_protocol=ATProto.LABEL, **record_kwarg)
|
|
|
|
|
create_task(queue='receive', obj=obj.key.urlsafe(), authed_as=op.repo)
|
2024-09-01 07:26:37 +00:00
|
|
|
# when running locally, comment out above and uncomment this
|
2024-09-01 14:46:15 +00:00
|
|
|
# logger.info(f'enqueuing receive task for {at_uri}')
|
2024-09-01 07:26:37 +00:00
|
|
|
except BaseException:
|
|
|
|
|
report_exception()
|
|
|
|
|
|
|
|
|
|
nonlocal seen
|
|
|
|
|
seen += 1
|
|
|
|
|
if limit is not None and seen >= limit:
|
|
|
|
|
return
|
|
|
|
|
|
2024-09-06 19:23:52 +00:00
|
|
|
last_stored_cursor = cur_timestamp = None
|
|
|
|
|
|
2024-09-01 14:46:15 +00:00
|
|
|
while frame := events.get():
|
2024-09-06 19:23:15 +00:00
|
|
|
header = libipld.decode_dag_cbor(frame)
|
|
|
|
|
# buf = BytesIO(frame)
|
2024-09-01 14:46:15 +00:00
|
|
|
|
|
|
|
|
# parse header
|
|
|
|
|
if header.get('op') == -1:
|
2024-09-06 19:23:15 +00:00
|
|
|
_, payload = libipld.decode_dag_cbor_multi(frame)
|
2024-09-01 14:46:15 +00:00
|
|
|
logger.warning(f'Got error from relay! {payload}')
|
|
|
|
|
continue
|
2024-09-05 22:20:35 +00:00
|
|
|
|
|
|
|
|
t = header.get('t')
|
|
|
|
|
if t != '#commit':
|
2024-09-06 19:23:15 +00:00
|
|
|
if t not in ('#account', '#identity', '#handle', '#tombstone'):
|
2024-09-05 22:20:35 +00:00
|
|
|
logger.info(f'Got {t} from relay')
|
2024-09-01 14:46:15 +00:00
|
|
|
continue
|
|
|
|
|
|
|
|
|
|
# parse payload
|
2024-09-06 19:23:15 +00:00
|
|
|
_, payload = libipld.decode_dag_cbor_multi(frame)
|
2024-09-01 14:46:15 +00:00
|
|
|
repo = payload.get('repo')
|
|
|
|
|
if not repo:
|
|
|
|
|
logger.warning(f'Payload missing repo! {payload}')
|
|
|
|
|
continue
|
|
|
|
|
|
|
|
|
|
seq = payload.get('seq')
|
|
|
|
|
if not seq:
|
|
|
|
|
logger.warning(f'Payload missing seq! {payload}')
|
|
|
|
|
continue
|
|
|
|
|
|
2024-09-06 19:23:52 +00:00
|
|
|
cur_timestamp = payload['time']
|
|
|
|
|
|
2024-09-01 14:46:15 +00:00
|
|
|
# if we fail processing this commit and raise an exception up to subscriber,
|
|
|
|
|
# skip it and start with the next commit when we're restarted
|
|
|
|
|
cursor.cursor = seq + 1
|
|
|
|
|
|
2024-09-06 19:23:52 +00:00
|
|
|
elapsed = util.now().replace(tzinfo=None) - cursor.updated
|
2024-09-01 14:46:15 +00:00
|
|
|
if (threading.current_thread().name == 'atproto_firehose.handler-0'
|
2024-09-06 19:23:52 +00:00
|
|
|
and elapsed > STORE_CURSOR_FREQ):
|
|
|
|
|
events_s = 0
|
|
|
|
|
if last_stored_cursor:
|
|
|
|
|
events_s = int((cursor.cursor - last_stored_cursor) /
|
|
|
|
|
elapsed.total_seconds())
|
|
|
|
|
last_stored_cursor = cursor.cursor
|
|
|
|
|
|
|
|
|
|
behind = util.now() - util.parse_iso8601(cur_timestamp)
|
|
|
|
|
|
|
|
|
|
# it's been long enough, update our stored cursor and metrics
|
|
|
|
|
logger.info(f'updating stored cursor to {cursor.cursor}, {events_s} events/s, {behind} ({int(behind.total_seconds())} s) behind')
|
2024-09-01 14:46:15 +00:00
|
|
|
cursor.put()
|
|
|
|
|
# when running locally, comment out put above and uncomment this
|
|
|
|
|
# cursor.updated = util.now().replace(tzinfo=None)
|
|
|
|
|
|
2024-09-01 07:26:37 +00:00
|
|
|
if payload['repo'] in bridged_dids:
|
|
|
|
|
logger.info(f'Ignoring record from our non-ATProto bridged user {payload["repo"]}')
|
2024-05-23 05:08:25 +00:00
|
|
|
continue
|
2024-05-07 21:15:51 +00:00
|
|
|
|
2024-05-08 20:26:36 +00:00
|
|
|
blocks = {}
|
2024-05-09 20:26:24 +00:00
|
|
|
if block_bytes := payload.get('blocks'):
|
2024-09-06 18:55:54 +00:00
|
|
|
# _, blocks = read_car(block_bytes)
|
|
|
|
|
# blocks = {block.cid: block for block in blocks}
|
|
|
|
|
_, blocks = libipld.decode_car(block_bytes)
|
2024-05-08 20:26:36 +00:00
|
|
|
|
2024-05-07 23:58:52 +00:00
|
|
|
# detect records that reference an ATProto user, eg replies, likes,
|
|
|
|
|
# reposts, mentions
|
2024-05-09 20:26:24 +00:00
|
|
|
for p_op in payload.get('ops', []):
|
2024-09-01 07:26:37 +00:00
|
|
|
op = Op(repo=payload['repo'], action=p_op.get('action'),
|
|
|
|
|
path=p_op.get('path'), seq=payload['seq'])
|
2024-05-09 20:26:24 +00:00
|
|
|
if not op.action or not op.path:
|
|
|
|
|
logger.info(
|
|
|
|
|
f'bad payload! seq {op.seq} has action {op.action} path {op.path}!')
|
|
|
|
|
continue
|
2024-05-08 20:26:36 +00:00
|
|
|
|
2024-09-01 07:26:37 +00:00
|
|
|
is_ours = op.repo in atproto_dids
|
2024-05-08 22:48:39 +00:00
|
|
|
if is_ours and op.action == 'delete':
|
|
|
|
|
logger.info(f'Got delete from our ATProto user: {op}')
|
2024-05-08 20:31:08 +00:00
|
|
|
# TODO: also detect deletes of records that *reference* our bridged
|
|
|
|
|
# users, eg a delete of a follow or like or repost of them.
|
|
|
|
|
# not easy because we need to getRecord the record to check
|
2024-09-01 07:26:37 +00:00
|
|
|
_handle(op)
|
2024-05-08 20:26:36 +00:00
|
|
|
continue
|
2024-05-07 23:58:52 +00:00
|
|
|
|
2024-05-09 20:26:24 +00:00
|
|
|
cid = p_op.get('cid')
|
2024-05-08 22:48:39 +00:00
|
|
|
block = blocks.get(cid)
|
2024-05-08 20:26:36 +00:00
|
|
|
# our own commits are sometimes missing the record
|
|
|
|
|
# https://github.com/snarfed/bridgy-fed/issues/1016
|
2024-05-09 20:26:24 +00:00
|
|
|
if not cid or not block:
|
2024-05-07 23:58:52 +00:00
|
|
|
continue
|
|
|
|
|
|
2024-05-19 06:17:07 +00:00
|
|
|
try:
|
2024-09-06 18:55:54 +00:00
|
|
|
# op = Op(*op[:-1], record=block)
|
|
|
|
|
op = Op(*op[:-1], record=block)
|
2024-05-19 06:17:07 +00:00
|
|
|
except BaseException:
|
|
|
|
|
# https://github.com/hashberg-io/dag-cbor/issues/14
|
|
|
|
|
logger.error(f"Couldn't decode block {cid} seq {op.seq}",
|
|
|
|
|
exc_info=True)
|
|
|
|
|
continue
|
|
|
|
|
|
2024-05-08 22:48:39 +00:00
|
|
|
type = op.record.get('$type')
|
2024-05-07 23:58:52 +00:00
|
|
|
if not type:
|
2024-05-08 22:48:39 +00:00
|
|
|
logger.warning('commit record missing $type! {op.action} {op.repo} {op.path} {cid}')
|
|
|
|
|
logger.warning(dag_json.encode(op.record).decode())
|
2024-05-07 23:58:52 +00:00
|
|
|
continue
|
|
|
|
|
|
2024-05-08 22:48:39 +00:00
|
|
|
if is_ours:
|
|
|
|
|
logger.info(f'Got one from our ATProto user: {op.action} {op.repo} {op.path}')
|
2024-09-01 07:26:37 +00:00
|
|
|
_handle(op)
|
2024-05-08 17:39:03 +00:00
|
|
|
continue
|
|
|
|
|
|
2024-05-07 23:58:52 +00:00
|
|
|
subjects = []
|
2024-05-08 18:18:50 +00:00
|
|
|
|
|
|
|
|
def maybe_add(did_or_ref):
|
|
|
|
|
if isinstance(did_or_ref, dict):
|
|
|
|
|
match = AT_URI_PATTERN.match(did_or_ref['uri'])
|
|
|
|
|
if match:
|
|
|
|
|
did = match.group('repo')
|
2024-05-08 18:31:31 +00:00
|
|
|
else:
|
|
|
|
|
return
|
2024-05-08 18:18:50 +00:00
|
|
|
else:
|
|
|
|
|
did = did_or_ref
|
|
|
|
|
|
2024-05-09 03:43:10 +00:00
|
|
|
if did and did in bridged_dids:
|
2024-05-07 23:58:52 +00:00
|
|
|
add(subjects, did)
|
|
|
|
|
|
|
|
|
|
if type in ('app.bsky.feed.like', 'app.bsky.feed.repost'):
|
2024-05-08 22:48:39 +00:00
|
|
|
maybe_add(op.record['subject'])
|
2024-05-07 23:58:52 +00:00
|
|
|
|
|
|
|
|
elif type in ('app.bsky.graph.block', 'app.bsky.graph.follow'):
|
2024-05-08 22:48:39 +00:00
|
|
|
maybe_add(op.record['subject'])
|
2024-05-07 23:58:52 +00:00
|
|
|
|
|
|
|
|
elif type == 'app.bsky.feed.post':
|
|
|
|
|
# replies
|
2024-05-08 22:48:39 +00:00
|
|
|
if reply := op.record.get('reply'):
|
2024-05-07 23:58:52 +00:00
|
|
|
for ref in 'parent', 'root':
|
2024-05-08 18:18:50 +00:00
|
|
|
maybe_add(reply[ref])
|
2024-05-07 23:58:52 +00:00
|
|
|
|
|
|
|
|
# mentions
|
2024-05-08 22:48:39 +00:00
|
|
|
for facet in op.record.get('facets', []):
|
2024-05-07 23:58:52 +00:00
|
|
|
for feature in facet['features']:
|
2024-05-08 18:31:31 +00:00
|
|
|
if feature['$type'] == 'app.bsky.richtext.facet#mention':
|
2024-05-07 23:58:52 +00:00
|
|
|
maybe_add(feature['did'])
|
|
|
|
|
|
2024-05-08 18:39:45 +00:00
|
|
|
# quote posts
|
2024-05-08 22:48:39 +00:00
|
|
|
if embed := op.record.get('embed'):
|
2024-05-08 19:13:26 +00:00
|
|
|
if embed['$type'] == 'app.bsky.embed.record':
|
2024-05-08 18:39:45 +00:00
|
|
|
maybe_add(embed['record'])
|
2024-05-08 19:13:26 +00:00
|
|
|
elif embed['$type'] == 'app.bsky.embed.recordWithMedia':
|
|
|
|
|
maybe_add(embed['record']['record'])
|
2024-05-07 23:58:52 +00:00
|
|
|
|
2024-05-08 17:39:03 +00:00
|
|
|
if subjects:
|
2024-05-08 22:48:39 +00:00
|
|
|
logger.info(f'Got one re our ATProto users {subjects}: {op.action} {op.repo} {op.path}')
|
2024-09-01 07:26:37 +00:00
|
|
|
_handle(op)
|
2024-05-08 17:39:03 +00:00
|
|
|
|
2024-05-09 13:43:51 +00:00
|
|
|
assert False, "handle thread shouldn't reach here!"
|