https://github.com/LemmyNet/lemmy/issues/3245

I posted far more details on the issue then I am putting here-

But, just to bring some math in- with the current full-mesh federation model, assuming 10,000 instances-

That will require nearly 50 million connections.

Each comment. Each vote. Each post, will have to be sent 50 million seperate times.

In the purposed hub-spoke model, We can reduce that by over 99%, so that each post/vote/comment/etc, only has to be sent 10,000 times (plus n*(n-1)/2 times, where n = number of hub servers).

The current full mesh architecture will not scale. I predict, exponential growth will continue to occur.

Let’s work on a solution to this problem together.

  • HTTP_404_NotFound@lemmyonline.comOP
    link
    fedilink
    English
    arrow-up
    6
    ·
    edit-2
    1 year ago

    for fucks sake, dude, be collaborative, and not defensive. This isn’t reddit, I am not out to attack your karma.

    If every instance, hosts a community, and Every other instance, subscribes to every one of those communities, that would lead to a full-mesh between all instances, resulting in worst-case scenario, ie, following the formula I provided for a full-mesh topology.

    That is indeed, the worst case scenario, I have provided, explained, and documented in my examples.

    • delcake@lemmy.songsforno.one
      link
      fedilink
      English
      arrow-up
      25
      ·
      1 year ago

      In no way is the person you’re responding to speaking defensively. They’ve discussed the reason why your extrapolation to a full-mesh connective worst-case scenario isn’t based in the reality of how ActivityPub functions. But you don’t seem to be willing to entertain the notion that the federation of any given action never exceeds the number of instances subscribed to the community that generated it.

      Even should every instance subscribe to every community on every other instance, the recipient of a federated action doesn’t turn around and rebroadcast that action back on to the network because it is not the authoritative host of that community. Therefore what this discussion is lacking is proof of where this exponential broadcast storm of federated actions comes from in your assertion.

    • King@vlemmy.net
      link
      fedilink
      English
      arrow-up
      14
      ·
      1 year ago

      Yes, it is a “full mesh” diagram. But for each specific “federated” action, it is a simple hub and spoke distribution. The hosting server will send the federated action to each subscribed node. The nodes don’t need to check in with each other for that specific action.

      I too believe that Federation is going to have scaling issues. But not due to full mesh

      • HTTP_404_NotFound@lemmyonline.comOP
        link
        fedilink
        English
        arrow-up
        2
        ·
        1 year ago

        I am onboard with you there-

        But, would not not agree- delegating and offloading those federation actions to a dedicated pool of servers, would not assist scalability?

        That way- each instance doesn’t need to maintain all of the connections?

        • King@vlemmy.net
          link
          fedilink
          English
          arrow-up
          5
          ·
          1 year ago

          There is no need to “maintain all of the connections”. The server opens a connection, sends the data, then closes the connection.

            • Fauxreigner@beehaw.org
              link
              fedilink
              English
              arrow-up
              6
              ·
              1 year ago

              Federation isn’t working well, but it’s not working well because the big instances aren’t able to keep up with all of the inbound/outbound messages, and if a message fails, that’s it. Right now there’s no automated way to resync and catch up on missed activity.

              • HTTP_404_NotFound@lemmyonline.comOP
                link
                fedilink
                English
                arrow-up
                1
                ·
                edit-2
                1 year ago

                So- what if, we can delegate a proxy/hub server, for managing all of the inbound/outbound messages, to offload that from the main instance server.

                ie, main instance sends/receives its messages through the proxy/hub server, the proxy/hub server then follows a pub/sub topology for sending and receiving.

                (Don’t imagine a centralized hub server, but, just imagine a localized proxy/hub server for your particular instance. Lets also assume, its designed where you can support multiple hub/proxy servers, in the event one gets overloaded)

                • Fauxreigner@beehaw.org
                  link
                  fedilink
                  English
                  arrow-up
                  1
                  ·
                  1 year ago

                  That doesn’t do anything to fix the problem. If a server can only handle 5k updates per minute (a completely made up number), it doesn’t matter if those 5k updates come from one server or a thousand. In theory you could cut down on outbound messages a bit if you could tell a “hub server” that post #123456 got another upvote, so please tell instances A, B, C, D, and E. But the total number of messages would increase, so even if the hub instance can handle more updates, it may eventually hit capacity again.

                  The core of the problem is that if an instance doesn’t process an update (inbound or outbound), it doesn’t ever retry, the instances are just out of sync for that post forever.

                  • HTTP_404_NotFound@lemmyonline.comOP
                    link
                    fedilink
                    English
                    arrow-up
                    1
                    ·
                    1 year ago

                    The core of the problem is that if an instance doesn’t process an update (inbound or outbound), it doesn’t ever retry, the instances are just out of sync for that post forever.

                    With the pub/sub method- that should be able to be minimized.

                    At least, with my experience of messing with rabbitmq- A message stays in the queue, until I have told rabbitMQ, Hey, I have processed this message.

                    If I accept a message, an encounter an exception mid-way through, that message returns back to the queue, until It has been processed, or dead-letter logic handles it.

                    Granted, there is a hard-coded timeout somewhere in lemmy, where, older messages cannot be processed. That would need to be adjusted.

              • cyd@vlemmy.net
                link
                fedilink
                English
                arrow-up
                1
                ·
                1 year ago

                How was syncing done in Usenet? It has a very similar decentralized model, and I don’t recall there being problems of data loss due to desyncing between servers.

            • King@vlemmy.net
              link
              fedilink
              English
              arrow-up
              1
              ·
              1 year ago

              I believe the current implementation wont scale because instances won’t be able to handle every subscribed federated action. Having a hub server doesn’t reduce the number of subscribed federated actions, only whom they come from.

              • HTTP_404_NotFound@lemmyonline.comOP
                link
                fedilink
                English
                arrow-up
                1
                ·
                1 year ago

                But- if we take that action of handling the federations, and seperate it from the main application server(Allowing the main instance server to focus on handling its local user-base), and architect it in a way that allows scaling the number of proxy servers up and down-

                Would that not sound like a big improvement to scalability?

                • King@vlemmy.net
                  link
                  fedilink
                  English
                  arrow-up
                  2
                  ·
                  1 year ago

                  The node still needs to receive every subscribed federated action and insert it into the local database. This has to be local to the “main application server”. Your proxy servers don’t reduce the number of federated actions. It only reduces the number of servers needed to communicate with.

                  I feel that the bottleneck will be the total number of federated actions, not which servers deliver them.

    • bdonvr@thelemmy.club
      link
      fedilink
      English
      arrow-up
      8
      ·
      edit-2
      1 year ago

      Apologies if I came off as hostile.

      I mean I get what you’re saying - I just don’t see the practical use. The centralized hub replication servers would have to basically foot a huge bill for the fediverse, and do so silently and invisibly to the end user. As it is, most instances run on goodwill or donations. A silent, invisible server is hard to gather donations for. Who would run them?

      Furthermore the topology you propose is essentially what we already have. A few large instances hold most of the largest communities. I don’t see that changing. This brings a fairly good balance - smaller instances pretty much only have to listen for updates from a few other instances, only the big instances are doing the hard work of notifying hundreds of others. They are already our “hubs”. Small instances really hardly do practically any hard work, the one I run for example just listens to maybe a dozen instances send updates, and occasionally sends out an update when one of my users interacts.

      I suppose I just don’t understand how this could be implemented in practice- or rather how it could be useful to do so. It would strictly enforce a sort of centralization that right now is only a natural consequence of user behavior, while seemingly only bringing theoretical benefits unlikely to be realized.

      • HTTP_404_NotFound@lemmyonline.comOP
        link
        fedilink
        English
        arrow-up
        2
        ·
        edit-2
        1 year ago

        The centralized hub replication servers would have to basically foot a huge bill for the fediverse, and do so silently and invisibly to the end user.

        One consideration, since they are only having to basically sub/pub - the load actually might be drastically lower than expected.

        Furthermore the topology you propose is essentially what we already have. A few large instances hold most of the largest communities. I don’t see that changing.

        Suppose- that is a valid point. The issue though- those large instances are unable to keep up with demand and load, causing lots of federation issues.

        Perhaps, my idea actually wouldn’t help that at all, but, using lemmy.ml as an example-

        Instead of it having to send all of its updates out to every server subscribed- it can delegate that to a hub server to do it. The hub server can run a very minimal set of instructions, with enough intelligence to handle sub/pub.

        Perhaps- one idea is, instead of thinking of it as a hub-server, think of it as a proxy server. Being able to delegate your instances actions to the proxy server to reduce that load from the main server.

        And, instead of the hubs/proxies being more centralized, perhaps, its just an optional thing which you CAN do.

        My line of thinking, is methods to reduce load from the main servers. This might be an idea that only benefits the handful of big servers.

        To also further clarify- I DONT have a solution to the problem. I am only intending to establish a forum to discuss if this is even a viable option, or perhaps, think of other ways to spread around the load.

        • monobot@lemmy.ml
          link
          fedilink
          English
          arrow-up
          2
          ·
          1 year ago

          I am not certain on scenarios you were mentioning above, but I do agree that separating software to instance plus hub/proxym/mssage queue could help with handling load.

          How can we scale our big i instances? I don’t know maybe it is easy to put instance on multiple servers, but sounds to me they are just buying bigger one, and that will fill up fast of growth continues to happen.

          I would like to hear from developers what they think, but thank you for starting conversation about scaling.

        • pe1uca@lemmy.pe1uca.dev
          link
          fedilink
          English
          arrow-up
          1
          ·
          1 year ago

          I’m just guessing here, but maybe a better solution would be to just to separate the flows inside the same server, this way the hosts of those server can use have containers dedicated to each flow.
          Also I don’t know much about architecture so take it with a grain of salt and correct me if I’m saying something silly. We could separate the lemmy container to have an API and a federation container.
          The clients still connect only to the API, the API offloads the federation requests to be sent to the specific container which can be hosted in another machine.
          So we’ll basically have the API and a message broker for the federation request, both being managed by the same admin since each host of a community should be responsible to send the updates to the users/instances subscribed.

        • russjr08@outpost.zeuslink.net
          link
          fedilink
          English
          arrow-up
          1
          ·
          1 year ago

          The issue though- those large instances are unable to keep up with demand and load, causing lots of federation issues.

          I am probably missing something / being really oblivious (its been a long day…) but wouldn’t this same problem occur to the hub server in your model?

          Although thinking about it a bit more, I thought I recalled seeing one of the Lemmy devs mention that the biggest issue is the SQL queries that are ran for various actions (such as loading the front page) - if that is the case, I don’t know if this idea would help with that.

          The idea of a centralized hub server(s) also sounds like we’d be moving closer to the model of a centralized Reddit… But I guess in a way, the fact that larger instances exist in of itself poses the same issue?

          … I’m probably just rambling to myself at this point, however, I do think a message queue type of system for federating events would be a good idea, for the sake of recovering from send failures.