Bridgy pulls comments and likes from social networks back to your web site. You can also use it to publish your posts to those networks.
Python HTML Other
Fetching latest commit…
Cannot retrieve the latest commit at this time.
Permalink
Failed to load latest commit information.
docs fix readthedocs: once more, with feeling: use pypi's gdata! Dec 20, 2017
scripts instagram: nice error when we can't fetch their site for indieauth Jan 5, 2018
static add github! Feb 12, 2018
templates Convert markdown in about.html to HTML Feb 21, 2018
test github misc: add notifications to listen scopes, add unit test Feb 15, 2018
.gitignore consolidate .gitignore a bit with wildcards Feb 15, 2018
.gitmodules start to port to virtualenv and pip for dependencies Jun 21, 2015
README.md add github! Feb 12, 2018
admin.py jinja: port all user-facing templates Nov 26, 2017
app.py add github! Feb 12, 2018
app.yaml add github! Feb 12, 2018
appengine_config.py docs: if virtualenv isn't in local/, use $VIRTUAL_ENV Dec 20, 2017
beta_users.txt add jackysee to beta users Nov 29, 2017
blog_webmention.py remove __author__s Dec 27, 2017
blogger.py remove __author__s Dec 27, 2017
circle.yml circle: switch from app engine SDK to gcloud Jun 20, 2017
cloud_storage_lifecycle.json bump up datastore backup expiration in cloud storage from 30d to 90d Jun 5, 2016
cron.py remove __author__s Dec 27, 2017
cron.yaml disable monthly datastore backup Oct 10, 2017
domain_blacklist.txt un-blacklist blog.mahabali.me Nov 15, 2017
dos.yaml blacklist 91.200.12.58, it's a script kiddie looking for xmlrpc vulns Jul 13, 2016
facebook.py facebook: interpret error 190 subcode 467 as dead token Jan 29, 2018
facebook_test_live.py facebook_test_live: update for passing state to FB oauth directly Dec 19, 2017
flickr.py remove __author__s Dec 27, 2017
github.py github: enable emoji reactions in backfeed Feb 22, 2018
googleplus.py remove __author__s Dec 27, 2017
handlers.py add github! Feb 12, 2018
index.yaml add Newer responses link. for #524 Oct 29, 2015
indieauth_client_id finish indieauth signup for instagram scraping! for #603 Apr 3, 2016
instagram.py instagram: nice error when we can't fetch their site for indieauth Jan 5, 2018
keys.md5 first pass at medium support! outbound webmentions only. #506 Aug 14, 2016
logs.py move log handler to webutil so i can reuse it in bridgy fed Oct 26, 2017
mapreduces.py noop: remove unused imports, other minor tweaks Dec 4, 2014
medium.py jinja: port all user-facing templates Nov 26, 2017
models.py more github docs, other minor tweaks Feb 21, 2018
original_post_discovery.py OPD: don't fetch homepages or feeds that are too big Oct 13, 2017
publish.py temporary: email me when new github users sign up and publish Feb 13, 2018
queue.yaml discover: handle errors, retry up to 3x Apr 26, 2017
requirements.freeze.txt switch back to upstream google-api-python-client, v1.6.5 (yay!) Jan 18, 2018
requirements.txt switch back to upstream google-api-python-client, v1.6.5 (yay!) Jan 18, 2018
superfeedr.py superfeedr: fail fast on blog post URLs over 500 chars Dec 20, 2017
tasks.py discover: for twitter reply URLs, also enqueue the in-reply-to tweet Jan 23, 2018
tumblr.py remove __author__s Dec 27, 2017
twitter.py convert Publish.type_label property to dynamic method Feb 12, 2018
util.py add util.requests_post() that injects our user agent Feb 12, 2018
webmention.py temporary: email me when new github users sign up and publish Feb 13, 2018
wordpress_rest.py remove __author__s Dec 27, 2017

README.md

Bridgy Bridgy Circle CI Coverage Status

Got a web site? Want replies and likes from social networks? Want to tweet, post, and more? Bridgy is for you.

https://brid.gy/

Bridgy pulls comments and likes from social networks back to your web site. You can also use it to publish your posts to those networks. See the user docs for more details, or the developer docs if you want to contribute.

License: This project is placed in the public domain.

Development

You'll need the App Engine Python SDK version 1.9.15 or later (for vendor support) or the Google Cloud SDK (aka gcloud) with the gcloud-appengine-python and gcloud-appengine-python-extras components. Add it to your $PYTHONPATH, e.g. export PYTHONPATH=$PYTHONPATH:/usr/local/google_appengine, and then run:

virtualenv local
source local/bin/activate
pip install -r requirements.freeze.txt

# We install gdata in source mode, and App Engine doesn't follow .egg-link
# files, so add a symlink to it.
ln -s ../../../src/gdata/src/gdata local/lib/python2.7/site-packages/gdata
ln -s ../../../src/gdata/src/atom local/lib/python2.7/site-packages/atom

python -m unittest discover

The last command runs the unit tests. If you send a pull request, please include (or update) a test for the new functionality if possible!

If you hit an error during setup, check out the oauth-dropins Troubleshooting/FAQ section. For searchability, here are a handful of error messages that have solutions there:

bash: ./bin/easy_install: ...bad interpreter: No such file or directory

ImportError: cannot import name certs

ImportError: No module named dev_appserver

ImportError: cannot import name tweepy

File ".../site-packages/tweepy/auth.py", line 68, in _get_request_token
  raise TweepError(e)
TweepError: must be _socket.socket, not socket

error: option --home not recognized

There's a good chance you'll need to make changes to granary, oauth-dropins, or webmention-tools at the same time as bridgy. To do that, clone their repos elsewhere, then install them in "source" mode with:

pip uninstall -y oauth-dropins
pip install -e <path to oauth-dropins>
ln -s <path to oauth-dropins>/oauth_dropins \
  local/lib/python2.7/site-packages/oauth_dropins

pip uninstall -y granary
pip install -e <path to granary>
ln -s <path to granary>/granary \
  local/lib/python2.7/site-packages/granary

pip uninstall -y webmentiontools
# webmention-tools isn't in pypi
ln -s <path to webmention-tools>/webmentiontools \
  local/lib/python2.7/site-packages/webmentiontools

The symlinks are necessary because App Engine's vendor module evidently doesn't follow .egg-link or .pth files. :/

To deploy to App Engine, run scripts/deploy.sh.

remote_api_shell is a useful interactive Python shell that can interact with the production app's datastore, memcache, etc. To use it, create a service account and download its JSON credentials, put it somewhere safe, and put its path in your GOOGLE_APPLICATION_CREDENTIALS environment variable.

Adding a new silo

So you want to add a new silo? Maybe MySpace, or Friendster, or even Tinder? Great! Here are the steps to do it. It looks like a lot, but it's not that bad, honest.

  1. Find the silo's API docs and check that it can do what Bridgy needs. At minimum, it should be able to get a user's posts and their comments, likes, and reposts, depending on which of those the silo supports. If you want publish support, it should also be able to create posts, comments, likes, reposts, and/or RSVPs.
  2. Fork and clone this repo.
  3. Create an app (aka client) in the silo's developer console, grab your app's id (aka key) and secret, put them into new local files in the repo root dir, following this pattern. You'll eventually want to send them to @snarfed and @kylewm too, but no hurry.
  4. Add the silo to oauth-dropins if it's not already there:
    1. Add a new .py file for your silo with an auth model and handler classes. Follow the existing examples.
    2. Add a button image.
    3. Add it to the app front page and the README.
  5. Add the silo to granary:
    1. Add a new .py file for your silo. Follow the existing examples. At minimum, you'll need to implement get_activities_response and convert your silo's API data to ActivityStreams.
    2. Add a new unit test file and write some tests!
    3. Add it to api.py (specifically Handler.get), app.py, app.yaml, index.html, and the README.
  6. Add the silo to Bridgy:
    1. Add a new .py file for your silo with a model class. Follow the existing examples.
    2. Add it to app.py, app.yaml, and handlers.py (just import the module).
    3. Add a 48x48 PNG icon to static/.
    4. Add new SILO_signup.html and SILO_user.html files in templates/ and add the silo to listen_signup.html. Follow the existing examples.
    5. Add the silo to about.html and this README.
    6. If users' profile picture URLs can change, add a cron job that updates them to cron.py and cron.yaml. Also add the model class to the datastore backup job there.
  7. Optionally add publish support:
    1. Implement create and preview_create for the silo in granary.
    2. Add the silo to publish.py: import its module, add it to SOURCES, and update this error message.
    3. Add a publish-signup block to SILO_user.html and add the silo to social_user.html.
    4. Update app.yaml.

Good luck, and happy hacking!

Monitoring

App Engine's built in dashboard and log browser are pretty good for interactive monitoring and debugging.

For alerting, we've set up Google Cloud Monitoring (née Stackdriver). Background in issue 377. It sends alerts by email and SMS when HTTP 4xx responses average >.1qps or 5xx >.05qps, latency averages >15s, or instance count averages >5 over the last 15m window.

Stats

I occasionally generate stats and graphs of usage and growth from the BigQuery dataset (#715). Here's how.

  1. Back up the full datastore to Google Cloud Storage. Include all entities except *Auth and other internal details.
    TODO: try the export service! It worked ok on 2018-01-01 but BigQuery choked on importing it with query: Entity was of unexpected kind "Response". (error code: invalidQuery).

  2. Import it into BigQuery:

    gsutil gs://brid-gy.appspot.com/
    # find the hash in the backup files, replace it in the paths below.
    
    for kind in BlogPost BlogWebmention Publish Response SyndicatedPost); do
      bq load --replace --source_format=DATASTORE_BACKUP datastore.$kind gs://brid-gy.appspot.com/aglzfmJyaWQtZ3lyQQsSHF9BRV9EYXRhc3RvcmVBZG1pbl9PcGVyYXRpb24Y-Z6kCAwLEhZfQUVfQmFja3VwX0luZm9ybWF0aW9uGAEM.$kind.backup_info
    done
    
    for kind in Blogger FacebookPage Flickr GooglePlusPage Instagram Medium Tumblr Twitter WordPress; do
      bq load --replace --source_format=DATASTORE_BACKUP sources.$kind gs://brid-gy.appspot.com/aglzfmJyaWQtZ3lyQQsSHF9BRV9EYXRhc3RvcmVBZG1pbl9PcGVyYXRpb24Y-Z6kCAwLEhZfQUVfQmFja3VwX0luZm9ybWF0aW9uGAEM.$kind.backup_info
    done
    
  3. Run the full stats BigQuery query. Download the results as CSV.

  4. Open the stats spreadsheet. Import the CSV, replacing the data sheet.

  5. Check out the graphs! Save full size images with OS screenshots, thumbnails with the Save Image button. Then post them!

Misc

The datastore is automatically backed up by a cron job that runs Datastore Admin backup and stores the results in Cloud Storage, in the brid-gy.appspot.com bucket. It backs up all entities monthly, and all entities except Response and SyndicatedPost weekly, since they make up 92% of all entities by size and they aren't as critical to keep.

We use this command to set a Cloud Storage lifecycle policy on that bucket that prunes older backups:

gsutil lifecycle set cloud_storage_lifecycle.json gs://brid-gy.appspot.com

Run this to see how much space we're currently using:

gsutil du -hsc gs://brid-gy.appspot.com/\*

Run this to download a single complete backup:

gsutil -m cp -r gs://brid-gy.appspot.com/weekly/datastore_backup_full_YYYY_MM_DD_\* .

Also see the BigQuery dataset (#715).