Last checked: 40 seconds ago
Get notified about any outages, downtime or incidents for RadarFirst and 1800+ other cloud vendors. Monitor 10 companies, for free.
Outage and incident data over the last 30 days for RadarFirst.
Join OutLogger to be notified when any of your vendors or the components you use experience an outage. It's completely free and takes less than 2 minutes!
Sign Up NowOutlogger tracks the status of these components for Xero:
Component | Status |
---|---|
RadarFirst Corporate Website | Active |
www.radarfirst.com | Active |
Radar Platform | Active |
api.radarfirst.com | Active |
app.radarfirst.com | Active |
Third Party | Active |
AWS ec2-us-west-2 | Active |
AWS rds-us-west-2 | Active |
View the latest incidents for RadarFirst and check for official updates:
Description: There was a brief gap this morning with our corporate website hosted on HubSpot as a new SSL certificate was being deployed and the existing SSL certificate was being revoked. During this gap, users would see in their web browser that the website was not secured with an active SSL certificate. We have since rectified the situation and have deployed the new active SSL certificate. We apologize for any inconvenience or concern this may have caused.
Status: Resolved
Impact: Minor | Started At: Feb. 5, 2019, 10:35 p.m.
Description: RADAR immediately asked for a root cause on i\).. how the issue was caused and ii\) why it took so long to fix. HubSpot responded with the following: HubSpot spent some time investigating the flow of events for RADAR's particular case. HubSpot uses JIRA to track engineering issues \(filed by customer support\). HubSpot recently rolled out a new "Customer Urgency" system to flag relevant and painful issues to the owning engineering leads of a particular area of the product. In this scenario, a JIRA ticket was actually **never** filed. Rather, a series of Slack messages were sent trying to find the correct people/teams to look into RADAR's urgent issue. This is actually against HubSpot’s own internal protocol and is likely one of the driving reasons why the issue took so long to get to the correct people. For obvious reasons, individual messaging is less than ideal, as conversations on progress and the issue itself were siloed in 1:1 conversations, so as people went in an out of meetings, etc. the full picture was lost and not put in front of the right people. If the JIRA ticket had been created with appropriate details and urgency, it likely would have been addressed much quicker, as everyone internally would have had visibility to the scope and details of the issue, and the engineering owners of the issue would have been flagged to address the critical issue. In addition, when the request to update the TLS settings for RADAR was selected for work, the operational team did not verify that RADAR had a custom domain and this information was required to be added to the certificate and wasn’t. HubSpot will be ensuring feedback is given to the involved parties, as well as make sure our "critical" protocol internal instructions properly reflect the correct protocol to ensure this support flow does not occur again.
Status: Postmortem
Impact: Major | Started At: Dec. 17, 2018, noon
Description: RADAR immediately asked for a root cause on i\).. how the issue was caused and ii\) why it took so long to fix. HubSpot responded with the following: HubSpot spent some time investigating the flow of events for RADAR's particular case. HubSpot uses JIRA to track engineering issues \(filed by customer support\). HubSpot recently rolled out a new "Customer Urgency" system to flag relevant and painful issues to the owning engineering leads of a particular area of the product. In this scenario, a JIRA ticket was actually **never** filed. Rather, a series of Slack messages were sent trying to find the correct people/teams to look into RADAR's urgent issue. This is actually against HubSpot’s own internal protocol and is likely one of the driving reasons why the issue took so long to get to the correct people. For obvious reasons, individual messaging is less than ideal, as conversations on progress and the issue itself were siloed in 1:1 conversations, so as people went in an out of meetings, etc. the full picture was lost and not put in front of the right people. If the JIRA ticket had been created with appropriate details and urgency, it likely would have been addressed much quicker, as everyone internally would have had visibility to the scope and details of the issue, and the engineering owners of the issue would have been flagged to address the critical issue. In addition, when the request to update the TLS settings for RADAR was selected for work, the operational team did not verify that RADAR had a custom domain and this information was required to be added to the certificate and wasn’t. HubSpot will be ensuring feedback is given to the involved parties, as well as make sure our "critical" protocol internal instructions properly reflect the correct protocol to ensure this support flow does not occur again.
Status: Postmortem
Impact: Major | Started At: Dec. 17, 2018, noon
Description: We had an unhealthy API instance, that wasn't detected by our Amazon EC2 health checks. The issue was intermittent because some requests were hitting one of the healthy API instances. We removed the unhealthy instance manually from the autoscaling group and it has since been replaced with a healthy instance. All services are back to full function. Ticket RADDEV-13610.
Status: Resolved
Impact: Minor | Started At: Dec. 14, 2018, 6:05 p.m.
Description: On June 14th starting at 5:10am customers started to experience 500 level errors when creating or updating incidents. During this time, users could still view incidents, report on incidents and perform other work with other parts of the application including the law library; no other functionality was impacted except the ability to create or update existing incidents. Customer support was contacted approximately at 5:30am and after troubleshooting escalated to operations team around 6:35am. Operations team determined it was caused by a known, but obscure issue, with Postgres 9.6.1. RADAR was put into maintenance mode at 7:50am and the database was patched. RADAR was placed back online at 8:06am with full services restored.
Status: Resolved
Impact: Major | Started At: June 14, 2018, 12:10 p.m.
Join OutLogger to be notified when any of your vendors or the components you use experience an outage or down time. Join for free - no credit card required.