cross-posted from: https://lemmy.world/post/2357075

It seems that self hosting, for oneself, a federated service, like Lemmy, would only serve to increase the traffic in the network, and not actually serve the purpose of load balancing between servers.

As far as I understand it, the way federation is supposed to work is that the servers cache all the content locally to then serve to the people that are registered to that server. In doing so, the servers only have to transmit a minimal amount of data between themselves which lowers the overhead for small servers – this then means that a small server doesn’t get overwhelmed by a ton of people requesting from it. Now, if, instead, you have everyone self hosting their own server, you go right back to having everyone sending a ton of requests to small servers, thereby overwhelming them. It seems that it’s really only beneficial to the network if you have, say, hundreds of medium sized servers instead of, say, thousands, of very small servers. While there is the resilience factor, the overhead of the network would be rather overwhelming.

Perhaps one possibility of fixing this is to use some form of load balancer like IPFS to distribute the requests more evenly, but I am no where even remotely close to being knowledgeable enough in that to say anything definitively.

  • treefingers@lemmy.world
    link
    fedilink
    English
    arrow-up
    5
    ·
    1 year ago

    Why is that?

    I’m not super familiar with torrenting protocols, but would have naively assumed that the very fact that subs have a single source of truth (e.g. selfhosted@lemmy.world is hosted on lemmy.world in its entirety, and then only cached on other lemmy instances) would be enough?

    I guess we’d need to federate the sub list, we wouldn’t want a central source of truth for that, but that bit isn’t any different to what we have currently AFAIK

    • shnizmuffin@lemmy.inbutts.lol
      link
      fedilink
      English
      arrow-up
      6
      arrow-down
      1
      ·
      1 year ago

      When you initially start a torrent, you define what “100%” is - all of the files. When you update a torrent, you need all of the updates. The beauty of a federated network is that the network can persist without all of it being available.

      I run my own instance. If every other server on the planet crapped out overnight, my instance would still be operable (with whatever content from the federation that I’ve consumed).

      The Fediverse is currently decentralized not distributed, and it should most definitely stay that way, for the sake of my disk space.

      • Aux@lemmy.world
        link
        fedilink
        English
        arrow-up
        4
        ·
        1 year ago

        Torrents are both decentralised and distributed.

        When you start a torrent, you don’t define a 100%, you define only your torrent and nothing else.

        To follow your example, if you run your own torrent instance and the network goes down, then of all torrents out there you will have whatever your instance managed to download. It works the exact same way in this regard.

        The main issue with decentralised P2P systems is that they’re very slow when user count is low.