• cadekat@pawb.social
    link
    fedilink
    arrow-up
    22
    ·
    1 year ago

    Won’t that lead to some horrible hug-of-death type scenarios if a post from a small instance gets popular on a huge one?

    • CoderKat@lemm.ee
      link
      fedilink
      English
      arrow-up
      42
      ·
      1 year ago

      Yes, but arguably it was never very scalable for federated software to store large media. It gets utterly massive quick. Third party image/video hosts that specialize in hosting those things can do a better job. And honestly, that’s the kinda data that is just better suited for centralization. Many people can afford to spin up a server that mostly just stores text and deals with basic interactions. Large images or streaming video gets expensive fast, especially if the site were to ever get even remotely close to reddit levels.

      • cadekat@pawb.social
        link
        fedilink
        arrow-up
        8
        ·
        1 year ago

        If you’re only responsible for caching for your own users, you don’t unduly burden smaller instances.

      • 30p87@feddit.de
        link
        fedilink
        arrow-up
        3
        arrow-down
        1
        ·
        1 year ago

        How would one realize CSAM protection? You’d need actual ML to check for it, and I do not think there are trained models available. And now find someone that wants to train such a model, somehow. Also, running an ML model would be quite expensive in energy and hardware.

        • NightAuthor@beehaw.org
          link
          fedilink
          English
          arrow-up
          2
          ·
          1 year ago

          There are models for detecting adult material, idk how well they’d work on CSAM though. Additionally, there exists a hash identification system for known images, idk if it’s available to the public, but I know apple has it.

          Idk, but we gotta figure out something