NovaCloud-Hosting - Notice history

Website - Operational

100% - uptime
Jul 2025 · 100.0%Aug · 100.0%Sep · 100.0%
Jul 2025
Aug 2025
Sep 2025

Dashboard - Operational

100% - uptime
Jul 2025 · 100.0%Aug · 100.0%Sep · 100.0%
Jul 2025
Aug 2025
Sep 2025
Operational

Dedicated-Server EYG1 - Operational

Dedicated-Server Management - Operational

FFM1 IP-Transit - Operational

100% - uptime
Jul 2025 · 99.55%Aug · 100.0%Sep · 100.0%
Jul 2025
Aug 2025
Sep 2025

Notice history

Sep 2025

No notices reported this month

Aug 2025

AMD Ryzen vHosts network connectivity issues
  • Postmortem
    Postmortem

    Postmortem: AMD Ryzen vHosts Network Connectivity Issues

    Incident Start: 31 August 2025, 9:00 PM CEST
    Incident End: 31 August 2025, 9:22 PM CEST
    Duration: Approximately 22 minutes
    Impact: Major outage on Ryzen vHosts, particularly RYZEN-05 and RYZEN-08

    Summary

    On 31 August 2025, a router misconfiguration caused a significant network outage affecting multiple Ryzen-based vHosts and services. The issue began at 9:00 PM CEST and was fully resolved by 9:22 PM CEST.

    Impact

    • RYZEN-05 and RYZEN-08 experienced full network loss for about 15 minutes.

    • All other Ryzen vHosts were briefly disconnected for approximately 10 seconds during the network recovery process.

    • Affected services included:

      • DB01 - Game-Cloud Infrastructure

      • Ryzen Game-Server Nodes (EYG1)

      • AMD Ryzen/High-End Root servers

    All customers with services on RYZEN-05 and RYZEN-08 are eligible to receive 2 days of additional runtime credit. Please open a support ticket to request this credit.

    Root Cause

    A router misconfiguration combined with an automation error caused connectivity issues on RYZEN-05 and RYZEN-08. Full restoration required reapplying routing configuration across all nodes, resulting in a brief interruption to all Ryzen vHosts.

    Resolution

    • A temporary automation fix was deployed to prevent recurrence of this specific issue.

    • Manual configuration changes restored full network connectivity by 9:22 PM CEST.

    Preventive Actions

    1. Implement per-node VM connectivity monitoring for faster detection and resolution.

    2. Improve automation workflows to avoid similar misconfigurations.

    3. Add validation checks to prevent faulty network configurations from being deployed.

    We apologize for the disruption and appreciate your patience as we implement additional safeguards to strengthen our network stability.

  • Resolved
    Resolved
    This incident has been resolved.
  • Identified
    Identified
    We are continuing to work on a fix for this incident.

Jul 2025

Multiple RYZEN VHOSTs + Dedicated servers unreachable
  • Resolved
    Resolved
    This incident has been resolved.
  • Postmortem
    Postmortem

    Summary,

    Between 20:29 and 20:32 GMT+2, several systems became temporarily unreachable due to a power distribution issue at the datacenter. The incident impacted specific power feeds and affected servers equipped with only a single power supply unit (PSU).

    Affected Systems,

    • Virtual Servers: RYZEN-VHOST 01, 03, 04, 05,

    • Dedicated Servers: A small number of systems with single-PSU configurations,

    • Unaffected: The majority of our infrastructure remained fully operational, including all dual-PSU systems,

    Root Cause,

    The incident was caused by a fault in one of the datacenter’s internal power subdistribution units. This resulted in a brief power loss on a specific feed.

    Resolution,

    Power was automatically restored by the datacenter within less than a minute. All affected systems came back online shortly afterward. We have confirmed that no data loss occurred.

    We are sorry for any inconvenience caused by this incident. Please open a ticket immediately if you still experience an outage, so we can address the issue.

  • Identified
    Identified

    There was a powerloss on one of the power feeds. It seems that the power has been restored within a few minutes and we are awaiting a statement from the datacenter to what has been affected and why.

    All affected Dedicated Servers and Root-Servers have already been started.

  • Investigating
    Investigating

    We are currently investigating this incident.

IP-Transit FFM1 - System replacement and upgrades
  • Completed
    23 July, 2025 at 9:59 PM
    Completed
    23 July, 2025 at 9:59 PM

    Maintenance has completed successfully.

    Since July 23, 2025, the setup has been successfully operating in active-active mode on both routers.

  • Update
    20 July, 2025 at 7:09 AM
    Update
    20 July, 2025 at 7:09 AM

    even if everything is currently running, the measures are not yet fully completed. we will leave this window open until all work is completed so that parallel operation of the routers works without any problems. at the moment, however, gretap, vxlan, wireguarda and gre should work until we continue later.

  • Update
    20 July, 2025 at 4:23 AM
    Update
    20 July, 2025 at 4:23 AM

    ok seems like that we have some issue going on after some runtime, looking into it...

  • Update
    20 July, 2025 at 4:04 AM
    Update
    20 July, 2025 at 4:04 AM

    since about 06:04 the load is already fully active on the new routers, please open a ticket if you encounter problems, currently everything seems to have worked as planned.

  • In progress
    20 July, 2025 at 4:00 AM
    In progress
    20 July, 2025 at 4:00 AM
    Maintenance is now in progress
  • Planned
    20 July, 2025 at 4:00 AM
    Planned
    20 July, 2025 at 4:00 AM

    We are planning for a scheduled maintenance during that time.

    This is our response to repeated minor interruptions since the migration from FRA1 to FFM1. Specifically, we will:

    - Increase the total processing bandwidth from 20 Gbps to 80 Gbps, i.e. quadruple it.

    - Replace the currently heavily loaded system with 2 new future-proof systems

    - Switchover to a BGP multipath setup for consistent redundancy and maximum reliability

    For affected customers, this means that there may be several outages during the maintenance-window until the work has been completed.

    We apologize for any inconvenience caused.

router1.ffm1 is down
  • Resolved
    Resolved
    is back up. This incident was automatically resolved by Instatus monitoring.
  • Postmortem
    Postmortem

    Summary

    On July 17th, 2025, starting around 05:30 GMT+2, we experienced a partial connectivity issue affecting traffic to the endpoint 5.83.150.1. A portion of external networks was unable to reach some prefixes including 5.83.150.0/24, while others remained fully functional.

    Initial observations pointed towards a potential BGP routing problem, but external route propagation appeared fully intact across all monitors.

    Root Cause

    The issue was caused by a malfunctioning SmartMitigate DDoS filtering node within the upstream provider's infrastructure. The affected host had stopped forwarding traffic.

    As a result, incoming traffic that was routed through this specific mitigation node was effectively dropped, leading to asymmetric reachability depending on which SmartMitigate host was assigned.

    This incident was not related to our own BGP announcements or configurations.

    Resolution

    The upstream provider identified the faulty SmartMitigate host and removed it from operation. Immediately afterward, traffic flows normalized and full connectivity to the affected prefix was restored. This occurred around 09:20 GMT+2.

    Impact Overview

    • Affected Prefix: 5.83.150.0/24

    • Root Cause: Faulty SmartMitigate (DDoS filter) node

    • BGP Status: Unaffected – all routes remained correctly announced and visible

    • Duration: ~3 hours 50 minutes

    • Impact: Partial traffic loss (limited to traffic routed via the affected SmartMitigate node)

    Current Status

    Connectivity has been fully restored. The root cause has been addressed at the upstream provider level. No further disruptions are expected related to this issue.

  • Identified
    Identified
    We are currently investigating a potential issue with our BGP announcements. Based on external reachability tests, it seems that a significant portion of the internet is unable to reach our endpoint, suggesting that traffic is not being properly propagated or accepted by certain upstream networks.
  • Investigating
    Investigating
    is down at the moment. This incident was automatically created by Instatus monitoring.

Jul 2025 to Sep 2025

Next