• jauntywundrkind 3 hours ago |
    A lot of this first law was specifically coupled to how these systems often hid that distributed objects were distributed. In the past 10 years, async has become far more common place, and it makes the distributed boundary much less like a secret special anomaly that you wouldn't otherwise deal with and far more like just another type of async code.

    I still thoroughly want to see capnproto or capnweb emerge the third party handoff, so we can do distributed systems where we tell microservice-b to use the results from microservice-a to run it's compute, without needing to proxy those results through ourself. Oh to dream.

    • teddyh an hour ago |
      Async fixes one problem with microservices. It does not fix the unexpected latency swings, the network timeouts and errors, the service disruptions when the microservice is unavailable, etc.
      • souvlakius 15 minutes ago |
        or the mismatch between request and response when using HTTP, or the overhead of using RPCs to protect against the previous scenario, or the issue of updating one microservice and not updating all the dependents
  • Havoc an hour ago |
    AI has also changed the dynamics around this. Splitting things into smaller components now has a dev advantage because the AI program better with smaller scope
    • mettamage an hour ago |
      Well yea... but the big con of microservices is still a thing: unexpected interactions

      But some of that could be mitigated I guess.

      • ori_b 18 minutes ago |
        Mitigating it is an unsolved research problem.
    • teddyh an hour ago |
      A separated component does not necessarily mean a microservice. It could be its own process, its own module, or even just its own function, which is fine. But microservices bring their own problems.
  • torginus an hour ago |
    > While small microservices are certainly simpler to reason about, I worry that this pushes complexity into the interconnections between services

    100% true in retrospect.

    • bandrami 44 minutes ago |
      I've found a lot of bugs in software in my career and basically none of them were at a single spot in a codebase. They've all been where two remote spots (or even more frequently to spots in two different codebases) interact. This was true even before microservices were a thing.
      • souvlakius 17 minutes ago |
        I mean, microservices split the code into smaller chunks but now lots of little pieces communicate over the network and unless you are using some form of RPC, this communication channels are not typed and there's a lot more stuff that could go wrong (packets dropped, DNS not resolving). Plus you could update one microservice and not update its dependents. I think a lot of people jumped on the hype without realising that it's a trade-off
      • kitd 17 minutes ago |
        There's a stat I think quoted in "Code Complete" by McConnell that says the number of bugs in a system strongly correlates with the number of coders. The conclusion is that as the # of coders goes up, the # of lines of communication between them grows exponentially, and it's the lines of (mis)communication that lead to bugs.

        This:

        1. explains Brooks' assertion that adding coders to a late project makes it later

        2. emphasises the importance in clearly defining interfaces between components, interfaces being the "paths of communication" between the coders of those components.

        So your assertion is well founded.

  • rednafi an hour ago |
    I dream of a SQL like engine for distributed systems where you can declaratively say "svc A uses the results of B & C where C depends on D."

    Then the engine would find the best way to resolve the graph and fetch the results. You could still add your imperative logic on top of the fetched results, but you don't concern yourself with the minutiae of resilience patterns and how to traverse the dependency graph.