Use cases
Software Products E-commerce MSPs Schools Development & Marketing DevOps Agencies Help Desk
Company
Internet Status Blog Pricing Log in Get started free

Outage in IONOS Cloud

AI Model Hub: Increased Error Rate

Resolved Minor
March 09, 2026 - Started about 2 months ago - Lasted 2 days
Official incident page

Incident Report

Summary AI Generated

IONOS Cloud's AI Model Hub experienced increased error rates with the Llama 405B model due to hardware degradation and subsequent capacity constraints following a hardware failure. The incident affected the Llama 3.1 405B Instruct model's performance and reliability over 58.9 hours. The service was restored with capacity constraints remaining, and users experiencing ongoing issues were advised to use GPT-OSS 120B as a temporary alternative while optimizations are deployed.

Our Model Hub Team is currently working on resolving errors related to an instance running the llama 405b model.

Trusted by 1,000+ teams

Never miss outages in third-party dependencies

Stop finding out about outages from your users. Monitor 6,320+ cloud services and get alerted the second something breaks.

IsDown status aggregator dashboard
Latest Updates ( sorted recent to last )
RESOLVED about 2 months ago - at 03/11/2026 07:20PM

We are marking this incident as resolved. The incident was caused by capacity constraints following a hardware failure. While capacity has been restored, we still see some usage‑specific constraints with the Llama 3.1 405B Instruct model. Our AI ModelHub team will deploy optimizations to the model to increase performance and reliability. We recommend that users still experiencing issues with the model check GPT‑OSS 120B as a potential (temporary) replacement.

MONITORING about 2 months ago - at 03/09/2026 06:53PM

Our AI Model Hub Team has mitigated the incident. While the underlying root cause is not yet fully established or resolved, the model service should be stable. We are monitoring the situation while the investigation is ongoing

IDENTIFIED about 2 months ago - at 03/09/2026 11:52AM

The team has identified the root cause: hardware degradation affecting this model's hosting environment is causing backend instability. We are currently implementing a fix.

INVESTIGATING about 2 months ago - at 03/09/2026 08:26AM

Our Model Hub Team is currently working on resolving errors related to an instance running the llama 405b model.

Never miss outages in third-party dependencies

With IsDown, you can monitor all your critical services' official status pages from one centralized dashboard and receive instant alerts the moment an outage is detected. Say goodbye to constantly checking multiple sites for updates and stay ahead of outages with IsDown.

Start free trial

No credit card required · Cancel anytime · 6320 services available

Integrations with Slack Microsoft Teams Google Chat Datadog PagerDuty Zapier Discord Webhook