• NeatNit@discuss.tchncs.de
    link
    fedilink
    arrow-up
    6
    ·
    edit-2
    7 hours ago

    I know a thing or two on how it actually works and I found the post funny. I know it doesn’t make sense but it’s still funny.

    Edit: to clarify (because it seems like you missed this point?), it’s about the recent downtime of AWS and of Cloudflare a few days later, each of which caused a huge portion of the internet to be inaccessible. The AWS downtime was caused by a DNS error (as ever), and I’m not sure about Cloudflare but it might be as well.

    • FauxLiving@lemmy.world
      link
      fedilink
      arrow-up
      3
      ·
      edit-2
      6 hours ago

      I’m not sure about Cloudflare but it might be as well.

      Cloudflare was a chain of unfortunate events.

      The TLDR is, a permission change caused a poorly written SQL query (without a properly filtering ‘where’ clause) to return a lot more data than normal. That data is used by an automated script to generate configuration files for the proxy services, because of the large return the configuration files were larger than normal (roughly 2x the size).

      The service that uses these configuration files has pre-allocated memory to hold the configuration files and the larger config file exceeded that size. This case, of having a file too large for the memory space, was improperly handled (ironically but not literally ironically, it was written in Rust) resulting in a thread panic which terminated the service and resulted in the 5xx errors.

      So, it’s more similar to the Crowdstrike crash (bad config file and poor error handling in a critical component).