bridgy-fed/README.md

6.6 KiB

Bridgy Fed Circle CI Coverage Status

Bridgy Fed connects different decentralized social network protocols. It currently supports the fediverse (eg Mastodon) via ActivityPub, Bluesky via the AT Protocol, and the IndieWeb via webmentions and microformats2. Farcaster and Nostr are under consideration. Bridgy Fed translates profiles, likes, reposts, mentions, follows, and more from any supported network to any other. See the user docs and developer docs for more details.

https://fed.brid.gy/

License: This project is placed in the public domain. You may also use it under the CC0 License.

Development

Development reference docs are at bridgy-fed.readthedocs.io. Pull requests are welcome! Feel free to ping me in #indieweb-dev with any questions.

First, fork and clone this repo. Then, install the Google Cloud SDK and run gcloud components install cloud-firestore-emulator to install the Firestore emulator. Once you have them, set up your environment by running these commands in the repo root directory:

gcloud config set project bridgy-federated
python3 -m venv local
source local/bin/activate
pip install -r requirements.txt

Now, run the tests to check that everything is set up ok:

gcloud emulators firestore start --host-port=:8089 --database-mode=datastore-mode < /dev/null >& /dev/null &
python3 -m unittest discover

Finally, run this in the repo root directory to start the web app locally:

GAE_ENV=localdev FLASK_ENV=development flask run -p 8080

If you send a pull request, please include (or update) a test for the new functionality!

If you hit an error during setup, check out the oauth-dropins Troubleshooting/FAQ section.

You may need to change granary, oauth-dropins, mf2util, or other dependencies as well as as Bridgy Fed. To do that, clone their repo locally, then install them in "source" mode with e.g.:

pip uninstall -y granary
pip install -e <path to granary>

To deploy to the production instance on App Engine - if @snarfed has added you as an owner - run:

gcloud -q beta app deploy --no-cache --project bridgy-federated *.yaml

How to add a new protocol

  1. Determine how you'll map the new protocol to other existing Bridgy Fed protocols, specifically identity, protocol inference, events, and operations. Add those to the existing tables in the docs in a PR. This is an important step before you start writing code.
  2. Implement the id and handle conversions in ids.py.
  3. If the new protocol uses a new data format - which is likely - add that format to granary in a new file with functions that convert to/from ActivityStreams 1 and tests. See nostr.py and test_nostr.py for examples.
  4. Implement the protocol in a new .py file as a subclass of both Protocol and User. Implement the send, fetch, serve, and target_for methods from Protocol and handle and web_url from User .
  5. TODO: add a new usage section to the docs for the new protocol.
  6. TODO: does the new protocol need any new UI or signup functionality? Unusual, but not impossible. Add that if necessary.
  7. Protocol logos may be emoji or image files. If this one is a file, add it static/. Then add the emoji or file <img> tag in the Protocol subclass's LOGO_HTML constant.

Stats

I occasionally generate stats and graphs of usage and growth via BigQuery, like I do with Bridgy. Here's how.

  1. Export the full datastore to Google Cloud Storage. Include all entities except MagicKey. Check to see if any new kinds have been added since the last time this command was run.

    gcloud datastore export --async gs://bridgy-federated.appspot.com/stats/ --kinds Follower,Object
    

    Note that --kinds is required. From the export docs:

    Data exported without specifying an entity filter cannot be loaded into BigQuery.

  2. Wait for it to be done with gcloud datastore operations list | grep done.

  3. Import it into BigQuery:

    for kind in Follower Object; do
      bq load --replace --nosync --source_format=DATASTORE_BACKUP datastore.$kind gs://bridgy-federated.appspot.com/stats/all_namespaces/kind_$kind/all_namespaces_kind_$kind.export_metadata
    done
    
  4. Check the jobs with bq ls -j, then wait for them with bq wait.

  5. Run the full stats BigQuery query. Download the results as CSV.

  6. Open the stats spreadsheet. Import the CSV, replacing the data sheet.

  7. Check out the graphs! Save full size images with OS or browser screenshots, thumbnails with the Download Chart button.