Bridgy connects your web site to social media. Likes, reposts, mentions, cross-posting, and more. See the user docs for more details, or the developer docs if you want to contribute.
Bridgy is part of the IndieWeb ecosystem. In IndieWeb terminology, Bridgy offers backfeed, POSSE, and webmention support as a service.
License: This project is placed in the public domain. You may also use it under the CC0 License.
Pull requests are welcome! Feel free to ping me in #indieweb-dev with any questions.
First, fork and clone this repo. Then, install the Google Cloud SDK and run gcloud components install cloud-firestore-emulator
to install the Firestore emulator. Once you have them, set up your environment by running these commands in the repo root directory:
gcloud config set project brid-gy
python3 -m venv local
source local/bin/activate
pip install -r requirements.txt
# needed to serve static files locally
ln -s local/lib/python3*/site-packages/oauth_dropins/static oauth_dropins_static
Now, you can fire up the gcloud emulator and run the tests:
gcloud emulators firestore start --host-port=:8089 --database-mode=datastore-mode < /dev/null >& /dev/null &
python3 -m unittest discover -s tests -t .
kill %1
If you send a pull request, please include or update a test for your new code!
To run the app locally, use flask run
:
gcloud emulators firestore start --host-port=:8089 --database-mode=datastore-mode < /dev/null >& /dev/null &
GAE_ENV=localdev FLASK_ENV=development flask run -p 8080
Open localhost:8080 and you should see the Bridgy home page!
To test a poll or propagate task, find the relevant Would add task line in the logs, eg:
INFO:root:Would add task: projects//locations/us-central1/queues/poll {'app_engine_http_request': {'http_method': 'POST', 'relative_uri': '/_ah/queue/poll', 'app_engine_routing': {'service': 'background'}, 'body': b'source_key=agNhcHByFgsSB1R3aXR0ZXIiCXNjaG5hcmZlZAw&last_polled=1970-01-01-00-00-00', 'headers': {'Content-Type': 'application/x-www-form-urlencoded'}}, 'schedule_time': seconds: 1591176072
...pull out the relative_uri
and body
, and then put them together in a curl
command against localhost:8080 (but don't run it yet!), eg:
curl -d 'source_key=agNhcHByFgsSB1R3aXR0ZXIiCXNjaG5hcmZlZAw&last_polled=1970-01-01-00-00-00' \
http://localhost:8080/_ah/queue/poll
Then, restart the app with FLASK_APP=background
to run the background task processing service, eg:
gcloud emulators firestore start --host-port=:8089 --database-mode=datastore-mode
GAE_ENV=localdev FLASK_ENV=development flask run -p 8080
Now, run the curl
command you constructed above.
If you hit an error during setup, check out the oauth-dropins Troubleshooting/FAQ section. For searchability, here are a handful of error messages that have solutions there:
bash: ./bin/easy_install: ...bad interpreter: No such file or directory
ImportError: cannot import name certs
ImportError: cannot import name tweepy
File ".../site-packages/tweepy/auth.py", line 68, in _get_request_token
raise TweepError(e)
TweepError: must be _socket.socket, not socket
error: option --home not recognized
There's a good chance you'll need to make changes to granary or oauth-dropins at the same time as bridgy. To do that, clone their repos elsewhere, then install them in "source" mode with:
pip uninstall -y oauth-dropins
pip install -e <path-to-oauth-dropins-repo>
ln -sf <path-to-oauth-dropins-repo>/oauth_dropins/static oauth_dropins_static
pip uninstall -y granary
pip install -e <path to granary>
To deploy to App Engine, run scripts/deploy.sh
.
remote_api_shell
is a useful interactive Python shell that can interact with the production app's datastore, memcache, etc. To use it, create a service account and download its JSON credentials, put it somewhere safe, and put its path in your GOOGLE_APPLICATION_CREDENTIALS
environment variable.
Deploying to your own App Engine project can be useful for testing, but is not recommended for production. To deploy to your own App Engine project, create a project on gcloud console and activate the Tasks API. Initialize the project on the command line using gcloud config set project <project-name>
followed by gcloud app create
. You will need to update TASKS_LOCATION
in util.py to match your project's location. Finally, you will need to add your "background" domain (eg background.YOUR-APP-NAME.appspot.com
) to OTHER_DOMAINS in util.py and set host_url
in tasks.py
to your base app url (eg app-dot-YOUR-APP-NAME.wn.r.appspot.com
). Finally, deploy (after testing) with gcloud -q beta app deploy --no-cache --project YOUR-APP-NAME *.yaml
To work on the browser extension:
cd browser-extension
npm install
npm run test
To run just one test:
npm run test -- -t 'part of test name'
If you're working on the browser extension, or you're sending in a bug report for it,, its JavaScript console logs are invaluable for debugging. Here's how to get them in Firefox:
about:debugging
Here's how to send them in with a bug report:
Here's how to cut a new release of the browser extension and publish it to addons.mozilla.org:
ln -fs manifest.firefox.json manifest.json
about:debugging
). Check that it works.browser-extension/manifest.json
.cd browser-extension/
npm test
./node_modules/web-ext/bin/web-ext.js build
Submit it to AMO.
# get API secret from Ryan if you don't have it
./node_modules/web-ext/bin/web-ext.js sign --api-key user:14645521:476 --api-secret ...
# If this succeeds, it will say:
...
Your add-on has been submitted for review. It passed validation but could not be automatically signed because this is a listed add-on.
FAIL
...
It's usually auto-approved within minutes. Check the public listing here.
Here's how to publish it to the Chrome Web Store:
ln -fs manifest.chrome.json manifest.json
chrome://extensions/
, Developer mode on). Check that it works.cd browser-extension/
npm test
./node_modules/web-ext/bin/web-ext.js build
browser-extension/web-ext-artifacts/
.0.7.0, 2024-01-03
0.6.1, 2022-09-18
0.6.0, 2022-09-17
0.5, 2022-07-21
0.4, 2022-01-30
0.3.5, 2021-03-04
0.3.4, 2021-02-22
0.3.3, 2021-02-20
0.3.2, 2021-02-18
0.3.1, 2021-02-17
0.2.1, 2021-01-09
0.2, 2021-01-03
0.1.5, 2020-12-25
So you want to add a new silo? Maybe MySpace, or Friendster, or even Tinder? Great! Here are the steps to do it. It looks like a lot, but it's not that bad, honest.
.py
file for your silo with an auth model and handler classes. Follow the existing examples.[NAME]_2x.png
, where [NAME]
is your start handler class's NAME
constant, eg 'twitter'
..py
file for your silo. Follow the existing examples. At minimum, you'll need to implement get_activities_response
and convert your silo's API data to ActivityStreams.api.py
(specifically Handler.get
), app.py
, index.html
, and the README..py
file for your silo with a model class. Follow the existing examples.app.py
and handlers.py
(just import the module).static/
.[SILO]_user.html
file in templates/
and add the silo to index.html
. Follow the existing examples.about.html
and this README.cron.py
.create
and preview_create
for the silo in granary.publish.py
: import its module, add it to SOURCES
, and update this error message.Good luck, and happy hacking!
App Engine's built in dashboard and log browser are pretty good for interactive monitoring and debugging.
For alerting, we've set up Google Cloud Monitoring (née Stackdriver). Background in issue 377. It sends alerts by email and SMS when HTTP 4xx responses average >.1qps or 5xx >.05qps, latency averages >15s, or instance count averages >5 over the last 15m window.
I occasionally generate stats and graphs of usage and growth from the BigQuery dataset (#715). Here's how.
Export the full datastore to Google Cloud Storage. Include all entities except *Auth
, Domain
and others with credentials or internal details. Check to see if any new kinds have been added since the last time this command was run.
gcloud datastore export --async gs://brid-gy.appspot.com/stats/ --kinds Activity,Blogger,BlogPost,BlogWebmention,Bluesky,Facebook,FacebookPage,Flickr,GitHub,GooglePlusPage,Instagram,Mastodon,Medium,Meetup,Publish,PublishedPage,Reddit,Response,SyndicatedPost,Tumblr,Twitter,WordPress
Note that --kinds
is required. From the export docs, Data exported without specifying an entity filter cannot be loaded into BigQuery. Also, expect this to cost around $10.
gcloud datastore operations list | grep done
or by watching the Datastore Import/Export page.for kind in Activity BlogPost BlogWebmention Publish SyndicatedPost; do
bq load --replace --nosync --source_format=DATASTORE_BACKUP datastore.$kind gs://brid-gy.appspot.com/stats/all_namespaces/kind_$kind/all_namespaces_kind_$kind.export_metadata
done
for kind in Blogger Bluesky Facebook FacebookPage Flickr GitHub GooglePlusPage Instagram Mastodon Medium Meetup Reddit Tumblr Twitter WordPress; do
bq load --replace --nosync --source_format=DATASTORE_BACKUP sources.$kind gs://brid-gy.appspot.com/stats/all_namespaces/kind_$kind/all_namespaces_kind_$kind.export_metadata
done
Open the Datastore entities page for the Response
kind, sorted by updated
ascending, and check out the first few rows: https://console.cloud.google.com/datastore/entities;kind=Response;ns=__$DEFAULT$__;sortCol=updated;sortDir=ASCENDING/query/kind?project=brid-gy
Open the existing Response
table in BigQuery: https://console.cloud.google.com/bigquery?project=brid-gy&ws=%211m10%211m4%214m3%211sbrid-gy%212sdatastore%213sResponse%211m4%211m3%211sbrid-gy%212sbquxjob_371f97c8_18131ff6e69%213sUS
Update the year in the queries below to three years before this year. Query for the same first few rows sorted by updated
ascending, check that they're the same:
SELECT * FROM `brid-gy.datastore.Response`
WHERE updated >= TIMESTAMP('202X-11-01T00:00:00Z')
ORDER BY updated ASC
LIMIT 10
Delete those rows:
DELETE FROM `brid-gy.datastore.Response`
WHERE updated >= TIMESTAMP('202X-11-01T00:00:00Z')
Load the new Response
entities into a temporary table:
bq load --replace=false --nosync --source_format=DATASTORE_BACKUP datastore.Response-new gs://brid-gy.appspot.com/stats/all_namespaces/kind_Response/all_namespaces_kind_Response.export_metadata
Append that table to the existing Response
table:
SELECT
leased_until,
original_posts,
type,
updated,
error,
sent,
skipped,
unsent,
created,
source,
status,
failed,
ARRAY(
SELECT STRUCT<`string` string, text string, provided string>(a, null, 'string')
FROM UNNEST(activities_json) as a
) AS activities_json,
IF(urls_to_activity IS NULL, NULL,
STRUCT<`string` string, text string, provided string>
(urls_to_activity, null, 'string')) AS urls_to_activity,
IF(response_json IS NULL, NULL,
STRUCT<`string` string, text string, provided string>
(response_json, null, 'string')) AS response_json,
ARRAY(
SELECT STRUCT<`string` string, text string, provided string>(x, null, 'string')
FROM UNNEST(old_response_jsons) as x
) AS old_response_jsons,
__key__,
__error__,
__has_error__
FROM `brid-gy.datastore.Response-new`
More => Query settings, Set a destination table for query results, dataset brid-gy.datastore, table Response, Append, check Allow large results, Save, Run.
Open sources.Facebook
, edit schema, add a url
field, string, nullable.
bq ls -j
, then wait for them with bq wait
.Final cleanup: delete the temporary Response-new
table.
Bridgy's online datastore only keeps responses for a year or two. I garbage collect (ie delete) older responses manually, generally just once a year when I generate statistics (above). All historical responses are kept in BigQuery for long term storage.
I use the Datastore Bulk Delete Dataflow template with a GQL query like this. (Update the years below to two years before today.)
SELECT * FROM Response WHERE updated < DATETIME('202X-11-01T00:00:00Z')
I either use the interactive web UI or this command line:
gcloud dataflow jobs run 'Delete Response datastore entities over 1y old'
--gcs-location gs://dataflow-templates-us-central1/latest/Datastore_to_Datastore_Delete
--region us-central1
--staging-location gs://brid-gy.appspot.com/tmp-datastore-delete
--parameters datastoreReadGqlQuery="SELECT * FROM `Response` WHERE updated < DATETIME('202X-11-01T00:00:00Z'),datastoreReadProjectId=brid-gy,datastoreDeleteProjectId=brid-gy"
Expect this to take at least a day or so.
Once it's done, update the stats constants in admin.py
.
The datastore is exported to BigQuery (#715) twice a year.
We use this command to set a Cloud Storage lifecycle policy on our buckets to prune older backups and other files:
gsutil lifecycle set cloud_storage_lifecycle.json gs://brid-gy.appspot.com
gsutil lifecycle set cloud_storage_lifecycle.json gs://brid-gy_cloudbuild
gsutil lifecycle set cloud_storage_lifecycle.json gs://staging.brid-gy.appspot.com
gsutil lifecycle set cloud_storage_lifecycle.json gs://us.artifacts.brid-gy.appspot.com
See how much space we're currently using in this dashboard. Run this to download a single complete backup:
gsutil -m cp -r gs://brid-gy.appspot.com/weekly/datastore_backup_full_YYYY_MM_DD_\* .