We had an issue with a core switch for our bladechassis that did not failover as it should following a fault.
Switch was still reporting functional and secondary switches did not take over as they should. This caused delay in troubleshooting process.
Effected datacenters: VBDC , interconnection between VBDC and HY
When: 12.55 CET – 13.41 CET
We’ll perform software updates during night time to prevent any software causes if this issue.
If it still happens again, we’ll replace the switch as faulty.
Adminor is scheduling a router maintenance effecting customers on IPranges: 184.108.40.206/20 and 220.127.116.11/22 .
Primary Service window is 2021-03-22 Monday 00.00 – 04.00 (two week notice!).
We are aiming for an outage less than 30min (15min or less usually), but keeping the window longer incase we hit an unforeseen snag or hardware issues that needs to be resolved.
Emergency service window (if we need to schedule sooner) 2021-03-15 Monday 00.00 – 04.00 CET.
We will try to keep this maintenance brief but are letting you know as our valued customer.
CloudFlare are having issues with their DNS services.
This may impact sites and services globally.
We’re starting to see some of their services returning to normal but we still caution that there may be more impacts.
More information at https://www.cloudflarestatus.com/
One of our network upstreams has announced network maintenance.
Please note that routing changes may cause momentary traffic dip during this time frame.
Our own traffic should automatically failover to secondary ISPs.
Date and time for maintenance window
Start date and time: 2020-05-05 03:00 UTC
End date and time: 2020-05-05 04:00 UTC
At 04.30 we performed a reboot of one of our core routers to ensure the issue is not on our end.
We will reboot some other equipment to troubleshoot issues.
We apologize for the disruption.
Update: One core switch was also rebooted manually at 05.50 – 05.52 . (short dip)
We have made adjustments to network traffic to find a stopgap solution to flapping network connection.
One of our upstream transits suffered an outage at 19.05. This caused network re-routing to happen and may have been felt as momentary network dip.
Traffic has been re-routed to failovers. We’re monitoring the networks.
When: 19:05 – 19:10 (upstream has failed over)
Impact: 5min re-routing, traffic dip
Update: Some more network dips occured during the evening. The network upstream suspected of causing high CPU usage in our cores have been isolated. We’re continuing to monitor the situation.
When: 20.30, 23.30 , duration (roughly 2-3min per incident)
At 22.00 we experienced some routing issues. We are currently investigating the cause of this.
Cause: One of our Upstream Transit lost BGP connectivity / flapped
Impact: 5 – 10min routing table rebuilds
Update (22.30): One of our transit operators having issues. Recovered after a few minutes.
At 04:00 we pushed an urgent patch to routers. This caused BGP to reset.
Traffic was disrupted momentarily until BGP was re-established.
We’ve applied important security updates to some of our switches in Västberga.
Each switch requires less than 5 minutes to reboot.
Impact: 08:50 – 08:55, 09:50 – 09.55
Problem: Upstream/ISP router rebooted. Suspected DDOS. Operator is restarting services and restoring connectivity.
Future resolution: Replacing upstream router.
We’ve now omitted the ISP with connectivity issues and re-routed traffic to different ISP.