The Twitter Engineering Blog

Information from Twitter's engineering team about our technology, tools and events.

Observability at Twitter: technical overview, part II

As one of the most critical infrastructure at Twitter, Observability provides highly scalable data collection and visualization services. This blog post gives overview of our architecture and shares our experience in developing and operating our systems.

Read more...

Observability at Twitter: technical overview, part I

As one of the most critical infrastructure at Twitter, observability provides highly scalable data collection and visualization services. Our post gives overview of our architecture and shares our experience in developing and operating our systems.

Read more...

Strong consistency in Manhattan

We explore lessons we learned while adding strong consistency to Manhattan and describe several problems that had to be solved along the way (implementing TTLs in a strongly consistent manner, doing distributed log truncations).

Read more...

When seconds really do matter

How we use secondly metrics to detect and resolve problems before they affect users.

Read more...

Power, minimal detectable effect, and bucket size estimation in A/B tests

Figuring out the minimal number of users one must expose to an experimental treatment to collect actionable data is not a trivial task. We explain how we approach this problem with Twitter’s A/B testing platform (DDG), and how we communicate issues of statistical power to experimenters.

Read more...

Fixing a recent password recovery issue

We recently learned about — and immediately fixed — a bug that affected our password recovery systems for about 24 hours last week. The bug had the potential to expose the email address and phone number associated with a small number of accounts (less than 10,000 active accounts). We’ve notified those account holders today, so if you weren’t notified, you weren’t affected.

Read more...

Distributed learning in Torch

We recently released Autograd for Torch, which greatly simplified our workflow when experimenting with complex deep learning architectures. The Twitter Cortex team is continuously investing in better tooling for manipulating our large datasets, and distributing training processes across machines in our cluster.

Today we’re open-sourcing four components of our training pipeline, so the community using Torch and/or Autograd can simplify their workflows when it comes to parallelizing training, and manipulating large, distributed datasets.

Read more...

Implications of use of multiple controls in an A/B test

Using a second control can be a tempting method of validating experiment results. We explore the statistics underlying usage of a second control, and conclude that this approach is strictly inferior to using a single large control.

Read more...

Visually explore funnels of user activities

We describe our experimental visual analytics approach for funnel analysis, which helps us explore how users interact with the user interfaces and gain new insights for improving user engagement with Twitter.

Read more...

Detecting and avoiding bucket imbalance in A/B tests

Some simple techniques to detect potentially biased implementations of A/B tests.

Read more...

Pages