NovaCloud-Hosting - Notice history

Website - Operational

100% - uptime
Nov 2024 · 99.87%Dec · 99.80%Jan 2025 · 99.99%
Nov 2024
Dec 2024
Jan 2025

Dashboard - Operational

100% - uptime
Nov 2024 · 99.87%Dec · 99.80%Jan 2025 · 99.99%
Nov 2024
Dec 2024
Jan 2025
Operational

Dedicated-Server FRA1 - Operational

Operational

Dedicated-Server EYG1 - Operational

Dedicated-Server Management - Operational

Notice history

Jan 2025

IP-Transit Maintenance
  • Update
    26 January, 2025 at 1:55 AM
    Completed
    26 January, 2025 at 1:55 AM

    After several days of intensive debugging and testing, we have come up with a rather ingenious workaround for the Wireguard-Bug. this means that no configuration changes are necessary on the customer side and the problem of wireguard simply changing the endpoint ip is finally history.

    thanks for your patience.

  • Completed
    25 January, 2025 at 8:33 PM
    Completed
    25 January, 2025 at 8:33 PM

    The tunnels have been running mostly stable again for some time and no further outages are to be expected.

    The interventions resulted in a downtime of approx. 38 minutes for IPv6 traffic, approx. 32 minutes for IPv4 traffic and approx. 1 hour for wireguard in general.

    VRRP is now stable again. In order to completely fix the Wireguard error, further measures are necessary on the customer side, they will be informed about this (it only affects a very small group of customers).

    As a result of the downtime, each customer has been allocated 3 days extra time, yes, this has already been done.

  • Update
    25 January, 2025 at 12:58 PM
    In progress
    25 January, 2025 at 12:58 PM

    it is currently stable, but not everything is implemented yet, but we will not make any other changes for the time being, as filters are unexpectedly in our way.

  • Update
    25 January, 2025 at 12:11 PM
    In progress
    25 January, 2025 at 12:11 PM

    the spin seems to be out, arp was filtered incorrectly. now only the persistence is checked.

  • Update
    25 January, 2025 at 12:05 PM
    In progress
    25 January, 2025 at 12:05 PM

    Unfortunately VRRP is still a bit bitchy, but it should be fixed soon.

  • In progress
    25 January, 2025 at 11:01 AM
    In progress
    25 January, 2025 at 11:01 AM

    We start with the implementation.

  • Planned
    25 January, 2025 at 11:00 AM
    Planned
    25 January, 2025 at 11:00 AM

    In the last 24+ hours we have noticed an error that VRRP is briefly interrupted during configuration reloading. this generates minimal packetloss.

    At the same time we are fixing a bug where some wireguard clients simply change the endpoint (which was set statically) to the secondary address of the routing servers when a VRRP “BACKUP” event occurs (e.g. because we need to migrate load or because a router fails or currently because of a config reload, which we also fix see above). This is quite rare but will be fixed now.

    This update is quite important to achieve a far-reaching stability and therefore planned for the short term.

    We apologize for the inconvenience, but we want to maximize uptime and stability for all with this.

    Short interruptions are to be expected during implementation.

Dec 2024

IP-Transit IPv6 connectivity issues
  • Resolved
    Resolved

    the problem will continue to be circumvented with the static ndp entry as it has been for the last few months. no further problems are to be expected. we will intensively test and fix the issue in the background. All affected customers who have either IPv6 endpoints OR IPv6 addresses are entitled to a 3-day extension of the runtime by means of a ticket. We apologize for the problems, this is not our claim.

  • Update
    Update

    the problem has reappeared. we have again added a permanent NDP entry which had also caused problems before. this should keep the problem stable for the time being. compensation in the form of runtime for affected customers is being clarified internally.

  • Update
    Update

    it seems to be stable again without any changes on our part. We have contacted the ISP because it has often attracted negative attention.

  • Identified
    Identified

    We are continuing to work on a fix for this incident. NDP is flapping.

  • Monitoring
    Monitoring

    The problem could be traced back to a strange problem with NDP which we will analyze further. the main instance is now back in operation and we implemented a fix for it and checking the stability now.

  • Update
    Update

    The problem occurs specifically only on our main instance (1). we maintain this instance and try to fix the problem. until then the traffic runs via the 2nd backup instance without interruption.

  • Update
    Update

    We (seems like that) found the cause and are implementing a fix. after each reload of the network config, a problem arose whereby the routers no longer had a functioning ipv6 for several minutes

  • Identified
    Identified
    We are continuing to work on a fix for this incident.

Nov 2024

Network Problems
  • Resolved
    Resolved

    This incident has been resolved.

    Update from ISP:

    „On 23.11.2024 at 23:12 (CET), our network experienced a sophisticated attack directly targeting our edge routing infrastructure. Due to the complexity of the attack, initial debugging efforts mistakenly concluded that we were dealing with a severe hardware failure within our edge routing equipment and that redundancy mechanisms were not functioning as expected. After manually triggering failovers to standby hardware in an attempt to rule out defective components without success, we intentionally disconnected our network from the internet to isolate and restore components one by one. This approach proved effective, with services beginning to recover at 23:41. IPv6 connectivity was restored quickly, while the last IPv4 prefixes came back online at 23:55. Unfortunately, the attackers quickly adjusted their methods, temporarily bypassing the newly implemented filters. As a result, another brief IPv4 outage occurred between 00:22 and 00:35, while IPv6 connectivity remained stable since the first fix was implemented.

    Since the incident occurred six hours ago, our team has been working through the night in collaboration with our upstream providers to implement permanent fixes against these new attack vectors. Although the network has faced further attacks in the meantime, it has remained stable, as our solutions have proven effective. Nevertheless, we remain on high alert over the coming hours and days to respond swiftly to any potential new attack patterns.

    This incident marks the first outage of our entire edge routing infrastructure, impacting every single dataforest customer, including IP transit services across all datacenters, since the launch of AS58212 nearly five years ago. Such an outage falls far short of the standards we set for ourselves, and we deeply apologize for the disruption caused to our customers. It is particularly frustrating that this attack succeeded despite our daily efforts to mitigate hundreds of similar attacks unnoticed. In this case, the attack managed to overwhelm our routers due to insufficient filtering against this unprecedented level of complexity.

    Please note that this incident was purely a reachability issue. There was no power outage, no hacking attempt, and no data breach.“

  • Monitoring
    Monitoring

    Our ISP has updated the incident:

    "Another fix has been implemented a few minutes ago and operation is stable since then. We are working on a permanent solution to avoid further outages. If any occur, we will report here."

  • Identified
    Identified
    We are continuing to work on a fix for this incident.
  • Update
    Update

    Our ISP has deescalated the incident.

    "We have implemented a fix. Network is stable for about five minutes now, IPv6 was not affected as much as IPv4 as remained available during the issue most of the time while IPv4 was affected heaviliy, leading to massive packet loss for a longer period of time. A post-mortem will follow tomorrow. Please rest assured we are 24/7 monitoring the status of our edge routing devices and will take action immediately if needed."

    We are sorry for the inconvenience caused by this rare accident.

  • Monitoring
    Monitoring

    The network is online at the moment, but we are waiting for the confirmation of our ISP.

  • Identified
    Identified

    Update from ISP:
    "We identified the issue and work on a solution"

  • Update
    Update

    Update from ISP:

    "We currently experience a outage in our edge routing, and investigating the situation."

  • Investigating
    Investigating

    We have detected a failure of our infrastructure in the main cube FRA01. The problem is already being investigated.

PayPal Zahlungsmethode unverfügbar
  • Resolved
    Resolved

    PayPal hat den Ausfall als abgeschlossen markiert und alle Probleme wurden behoben.

    Somit ist es wieder möglich ohne Probleme PayPal als Zahlungsmethode zu nutzen.

  • Identified
    Identified

    Die API ist teilweise wieder verfügbar und es ist möglich sich in PayPal Accounts anzumelden. Es kommt teilweise bei Zahlungen zu Fehlern. Falls ein Fehler auftreten sollte bitte einmal zurück gehen und es nach ein paar Minuten nochmals versuchen.

  • Investigating
    Investigating

    PayPal scheint ein Problem mit Accounts und der API zu haben, weshalb wir im Moment keine Zahlungen per PayPal bearbeiten können. Es ist generrell nicht möglich sich in PayPal accounts einzuloggen.

    Auf https://www.paypal-status.com/incident/production gibt es genauere Infos von PayPal direkt.

    Wir entschuldigen uns für umstände, jedoch haben wir auf solche Ausfälle keinen Einfluss.

EPYC-01-Vhost outage
  • Resolved
    Resolved

    Unsere Server sind soeben gestartet, alle Server sollten in den nächsten 10 Minuten online sein.

  • Investigating
    Investigating

    Das SkyLink Rechenzentrum hat anscheind einen Stromausfall erlitten, was aus mehreren Quellen bestätigt wurde, jedoch nicht offiziell von dem Betreiber selbst.

    Da es momentan in der Region einen Stromausfall gibt, wurde auf die USV-Anlagen umgeschalten, welche für kurze Zeit Strom geliefert haben. Es scheint Probleme mit den Dieselaggegraten zu geben, weshalb das gesamte Rechenzentrum offline ist.

    https://x.com/EnexisStoringen/status/1858846625310925068

    Wir vermuten eine verbesserung in den kommenden Stunden, können jedoch im Moment keine genauen Angaben dazu geben.


    The SkyLink data center appears to have suffered a power outage, as confirmed by multiple sources, though not officially by the operator itself.

    Currently, the region is experiencing a power outage, and the data center switched to UPS systems, which provided power for a short time. However, it seems there are issues with the diesel generators, resulting in the entire data center being offline.

    https://x.com/EnexisStoringen/status/1858846625310925068

    We anticipate improvements in the coming hours but cannot provide any specific estimates at the moment.

Nov 2024 to Jan 2025

Next