Buttondown

Buttondown Status

Official source
Everything seems OK

Don't miss the next Buttondown outage!

Get a notification when the status changes! Start your FREE account today!

Sign up with Google or with email

Easily monitor 30 service providers that your business relies on for only $7/month.
Monitor up to 5 services for free.

Stats

0 incidents in the last 7 days

0 incidents in the last 30 days

Automatic Checks

Last check: 4 minutes ago

Last known issue: 2 months ago

Latest Incidents

Last 30 days

15/04
16/04
17/04
18/04
19/04
20/04
21/04
22/04
23/04
24/04
25/04
26/04
27/04
28/04
29/04
30/04
01/05
02/05
03/05
04/05
05/05
06/05
07/05
08/05
09/05
10/05
11/05
12/05
13/05
14/05
15/05

Resolved DNS is not resolving in US-EAST

This is a DNSimple issue, unfortunately: https://dnsimple.statuspage.io/incidents/bcnzxs0mgz21=

Resolved General availability outage

For around an hour this morning, Buttondown had significantly degraded availability. ## What happened? New hosts refused to spin up and were "correctly" throwing 500s for around 30% of requests (this was only impacting hosts that were automatically cycling in and out, which is why it wasn't all requests.) ## Why did this happen? I'm using an undocumented Notion API to power documentation search, and the token that I was using to power that API expired in a way that I was not defensively programming against. This meant that each time the server tried to restart it would hit the API, fall over, and then pass that failure onto the client. As soon as this happened widespread enough, I got an alert for it... but I was out on a run. As soon as I got back, I hit the circuit breaker for that codepath and things got back to normal. ## Why won't this happen again? That circuit breaker is gonna stay off for a little, but I plan on moving all of that compilation to a build-time step anyway, removing the Notion codepath from the critical path of the application! ## Any questions? Email me: [email protected]

Resolved General availability outage

For around five hours (meaning the early morning of 11/13, Pacific time) Buttondown's availability was heavily degraded. ## What happened? Around 50-70% of requests timed out. It wasn't _quite_ a complete DDOS, but essentially so. ## Why did this happen? This is... actually fairly silly, as far as these things go. An old third-party log handler that Buttondown was using shut off access at the logdrain I was using. (This is a totally reasonable thing to do!) _Unfortunately_, that clobbered a huge amount of the requests being served, to the point where all the active dynos on my infrastructure were busy complaining and throwing errors because they couldn't emit logs. The irony of this does not escape me. ## Well, why did it take so long to fix? I was asleep. No, really! That's the reason. I've got two thresholds for Buttondown outages: 1. The server is down for a little, which texts me. 2. The server is hard down for all requests, which calls and pages me. This was an exceptionally long bout of the former, which meant I woke up to like seventy outage texts but no outright pages. ## Why won't this happen again? First: I've upped (or lowered, depending on how you look at it) the threshold for what constitutes an outage. I made a lot of these alerts two years ago when Buttondown was a fraction of a fraction of its current size; thankfully, things are generally stable, but its still time to be more alert. Any non-trivial breakage of traffic pages little old me. Second: to fix the _actual_ issue, I'm spending some time this weekend messing around with the logging & error infrastructure Buttondown uses to more gracefully degrade. Have any questions? Email me: [email protected]

Resolved Broken tracking links

I tried to switch over to SSL for our tracking links and it looks like something got awry. I need to follow up with CloudFlare as to what the issue was (it was likely a misconfiguration on my end, in all honesty, but hard to say) — but for around two hours all tracked links were broken because they pointed to https and not http. I've reached out to the folks who have sent outbound emails during that time, and am in the process of backfilling the transactional emails (such as subscriber confirmations) that were sent during that window!

12 months ago Official incident report

Resolved 500s on Frontend Application

GitHub is running into [issues](https://www.githubstatus.com/incidents/80d0cs6kpsps) and Buttondown's CI pipeline wasn't running properly, leading to a feature branch getting deployed to production which caused breakages for a few minutes.

about 1 year ago Official incident report

Incidents will happen.
You just need to prepare the best for them.

We're monitoring 806 services and adding more every week.

All-in-one platform to check the status of your services

IsDown integrates with hundreds of services. Handles the hassle of going to each one of the status pages and manage it one by one. We also help control how you receive the notifications.
We monitor all problems and outages and keep you posted on their current status in almost real-time.

Get notifications in your favorite communication channel

You can easily get notifications in your email, slack, or use Webhooks and Zapier to introduce the service status in your workflows.


Email
Slack
Slack
Zapier
Zapier

Webhooks

Empower your teams with more data

IsDown collects status data from services to help you be ahead of the game.

Engineering

You already monitor internal systems. Add another dimension (external systems) to your monitoring data and complement it with the external factors.

Customer Support

Know before your clients tell you. Anticipate possible issues and make the necessary arrangements.

Marketing

One of your competitors is down? Maybe a good time to spread the word about your service.

Trusted by teams from all over the world

Services Available
806
Incidents registered
47832
Monitoring Incidents
60

Ready to dive in? Start using IsDown today.

Sign up for free