I hope this won’t be counted as some form of self-promotion, even though I am sharing a post from my own blog.

As a tech worker who works in a Cloud shop, I wanted to elaborate the many reasons why I find working with Clouds terrible, from multiple points of view.

I tried to organize my thoughts in a (relatively long) post, in which both technical aspects and political aspects (which are very related) are covered.

I am sure many people will have different perspectives, and this could be potentially also a nice prompt for a discussion.

  • conciselyverbose@sh.itjust.works
    link
    fedilink
    English
    arrow-up
    37
    ·
    5 months ago

    So the whole thing is well worth a read IMO, and addresses a lot of the issues I have with cloud as the solution for everything.

    My main point here is that individuals and organizations that require all the flexibility that cloud services offer are a (tiny) minority. This means that for the majority of us, all the complexity necessary to provide this flexibility ends up being purely a complication or worse, a liability.

    There are absolutely companies who need the scaling. But it’s a fucking lot of overhead if you don’t.

    Let’s repeat it one more time: complexity hides and creates security issues.

    This is similar to all the LLM code stuff. If you don’t actually fully understand what your code does, bad stuff happens.

    This premise has the consequence that Cloud systems are a big puzzle. The pieces of the puzzle are the Cloud products. Engineers working with Cloud systems essentially need to understand the abstraction but not necessarily the underlying, ultimate working mechanism of what those abstractions do. For example, a cloud expert might know everything about the difference between NACLs and Security Groups, all the details about how to configure them, their limitations etc., but the main idea is that such expert doesn’t need to know anything below that (e.g., how the traffic is filtered).

    Ultimately my perspective, and I appreciate it’s a very personal one, is that building and working with the Cloud makes me feel like a glorified application administrator. My job becomes researching how the Cloud solved the problem that I need to solve, and compose the solution in the way the Cloud provider imagined it should be solved, rather than solving the problem

    I was going to bring up basically this point:

    because vendor-lock is not something that has only to do with infrastructure. It has also to do with the skills of the engineers involved. Cloud knowledge, for the most part, is not portable. You are a wizard of IAM policies in GCP? Good job, this is completely useless if you go to Azure. Oh, you are a guru of VPCs and private endpoints? Well done, this is completely useless if you move to a different cloud.

    But you covered it pretty well. Abstractions are great. Proprietary abstractions that are more focused on how they can bill you than real, useful, functional categories? Not so much.

    Despite the efforts means something which is ironic: many companies which run on Cloud, at some point, will have one or more teams whose main purpose is understanding how they are spending money in the Cloud and to reduce those costs. If this sounds conflicting with the idea of reducing personnel, well, it is. The digital infrastructure of my organization is not that huge. Give or take 2000 compute instances (some very small). Something that 200 servers could easily provide. Cloud bills are more than $15 millions/year. I checked a server builder for example, and an absolute beast (something like 2x Xeon platinum processor, 200TB of NVME disks, 1TB of RAM etc.) would still stay comfortably under $250k. 100 servers this powerful will probably be a multiple of our computing power, and cost almost a third if we consider a lifetime of 3 years, which is very low. A more realistic estimation of 5 years leads to a saving of ~$50 millions over 5 years. Completely insane! This is of course if you want to buy hardware. Powerful servers rented run you for $500-1000/month. Assuming a cost of $1000/month, my company could rent more than 1000 powerful servers, and still save money compared to Cloud costs, leaving plenty for additional services such as networking, storage, premium support (remote hands) or actual engineers salary

    So there’s a level of rent seeking behind all the software moving to subscriptions, and them wanting to lock you in just like their service providers are doing to them. But I have to think the massive costs of cloud junk also pay a role in stuff like a calendar charging double digit annual fees for something that takes very little storage and very little computation (and you of course can’t just buy software any more).

    I have no words for multi-cloud. Even like a Facebook or YouTube scale site, are you really going to double (or more for some reason?) your storage costs (plus whatever intercommunication between the two), just in case the provider goes down for a couple hours (which is extremely rare, and you won’t be the only site impacted, so people won’t really blame you for.) Plus that architecture sounds like the shitshow to end all shitshows.

    • BearOfaTime@lemm.ee
      link
      fedilink
      English
      arrow-up
      19
      ·
      edit-2
      5 months ago

      Agreed on it all.

      I think a big driver for cloud clients is bean counters - cloud is an expense, while having your own systems is capital investment.

      They’d rather have the waste of leasing too much compute than have to pay taxes on systems plus the cost of staff to run it.

      We won’t really see this get addressed until companies have to truly own the risks they take on (see all the hacks that happen on a daily basis because CIO won’t pay for the security that IT management is screaming to build). When fines for these breaches are meaningful, cloud will be less interesting.

    • loudwhisper@infosec.pubOP
      link
      fedilink
      English
      arrow-up
      15
      ·
      5 months ago

      Thanks!

      But I have to think the massive costs of cloud junk also pay a role in stuff like a calendar charging double digit annual fees for something that takes very little storage and very little computation (and you of course can’t just buy software any more).

      Absolutely agree. I did not even think about this aspect, but I think you are absolutely spot on. Building something with huge costs is something that ultimately gets passed to the users in addition to the rent-seeking aspect.

      I have no words for multi-cloud.

      You and me both. I have to work with it and the reality is, there is nobody who actually understands the whole thing. The level of complexity (and fragility, I might add) of it all is astonishing. And all of this to mitigate some (honestly) low risk of downtime from the cloud provider. I have lobbied a little bit against at work, but ultimately it has become a marketing tool to sell to customers, so goodbye any hope of rational evaluation…

      • Tja@programming.dev
        link
        fedilink
        English
        arrow-up
        1
        ·
        5 months ago

        It’s all shits and giggles until a network config takes down your cloud provider for 11 hours and you can’t even look at the logs. And multicloud is quite robust if done right, more so than a single cloud, if your setup is fragile someone is not doing their job right.

        • loudwhisper@infosec.pubOP
          link
          fedilink
          English
          arrow-up
          1
          ·
          5 months ago

          Complexity brings fragility. It’s not about doing the job right, is that “right” means having to deal with a level of complexity, a so high number of moving parts and configuration options, that the bar is set very high.

          Also, I would argue that a large number of organizations don’t actually need the resilience that they pay a very high price for.

          • Tja@programming.dev
            link
            fedilink
            English
            arrow-up
            3
            ·
            5 months ago

            Complexity in this case should bring redundancy, not fragility. You are adding components in parallel, not in series, thus reducing fragility.

            A raid 5 is more complex than a single drive, but it’s less fragile.

            • loudwhisper@infosec.pubOP
              link
              fedilink
              English
              arrow-up
              1
              ·
              5 months ago

              I wish it worked like that, but I donct think it does. Connecting clouds means introducing many complex problems. Data synchronization and avoiding split-brain scenarios, a network setup way more complex, stateful storage that needs to take into account all the quirks and peculiarities of all services across all clouds, service accounts and permissions that need to be granted and segregated for all of them, and way more. You may gain resilience in some areas, but you introduce a lot more things that can fail, be misconfigured or compromised.

              Plus, a complex setup makes it harder by definition to identify SPOFs, especially considering it’s very likely nobody in the workforce is going to be an expert in all the clouds in use.

              To keep using your simile of the disks, a single disk with a backup might be a better solution for many people, considering you otherwise might need a RAID controller that can fail and all the knowledge to handle and manage a RAID array properly, in addition to paying 4 or 5 times the storage. Obviously this is just to make a point, I don’t actually think that RAID 5 vs JBOD introduces comparable complexity compared to what multi-cloud architecture does to single-cloud.

              • Tja@programming.dev
                link
                fedilink
                English
                arrow-up
                1
                ·
                5 months ago

                Split brain are easily solved, there’s of the shelf solutions and if you have some custom code you can use plenty of well researched solutions, for instance raft. Putting bizantine fault in Google scholar yields thousands of papers,if you want something fancier.

                Same for most problems you mentioned, they were an issue 10 years ago, nowadays you can federate, abstract or outsource most of it.

                Making it harder to identify SPFOs doesn’t increase fragility. If you whole system a single instance it’s trivial to identify (the whole thing) but very brittle.

                • loudwhisper@infosec.pubOP
                  link
                  fedilink
                  English
                  arrow-up
                  1
                  ·
                  5 months ago

                  Of course the problem is solved, but that doesn’t mean that the solution is easy. Also, distributed protocols still need to work on top of a complicated network and with real-life constraints in terms of performances (to list a few). A bug, misconfiguration, oversight and you have a problem.

                  Just to make an example, I remember a Kafka cluster with 5 replicas completely shitting its pants for 6h to rebalance data during a planned maintenance where one node was brought offline. It caused one of the longest outages to date with the websites which relied on it offline. Was it our fault? Was it a misconfiguration? A bug? It doesn’t matter, it’s a complex system which was implemented and probably something was missed.

                  Technology is implemented by people, complexity increased the chances of mistakes, not sure this can be argued.

                  Making it harder to identify SPOF means you might miss your SPOF, and that means having liabilities, and having anyway scenarios where your system can crash, in addition for paying quite a lot to build a resilience that you don’t achieve.

                  A single instance with 2 failure scenarios (disk failure and network failure) - to make an example - is not more fragile than a distributed system with 20 failure scenarios. Failure scenarios and SPOF can have compensating controls and be mitigated successfully. A complex system where these can’t be fully identified can’t have compensating control and residual risk might be much harder. So yes, a single disk can fail more likely than 3 disks at once, but this doesn’t give the whole picture.

  • a Kendrick fan@lemmy.ml
    link
    fedilink
    English
    arrow-up
    25
    ·
    5 months ago

    Great post, a quick nitpick if you don’t mind, introduce or use an abbreviation’s full words before using its abbreviated form

    Granted that the article is geared towards sysadmins and cloud developers, others who may want to read it may have a hard time doing so. As an example, reading through the first technical point, I saw “IAMs” and “Network ACL”, I don’t understand what those abbrs mean

    • loudwhisper@infosec.pubOP
      link
      fedilink
      English
      arrow-up
      11
      ·
      edit-2
      5 months ago

      Thanks, that is a very good observation! I will try to sneak an edit later today where I can add some appendix about acronyms and abbreviations.

      Edit:

      While it might not look great, I have added at the bottom an Appendix with all (hopefully, I might have missed some) acronyms and abbreviations. Thanks for the suggestion!

  • Ledivin@lemmy.world
    link
    fedilink
    English
    arrow-up
    33
    arrow-down
    10
    ·
    edit-2
    5 months ago

    I’m sorry, but this started like a recipe article and I lost all interest. I don’t care about your life story, I clicked the link to read your opinions, and you spent the first several paragraphs avoiding them.

    • loudwhisper@infosec.pubOP
      link
      fedilink
      English
      arrow-up
      34
      arrow-down
      3
      ·
      5 months ago

      Nothing to be sorry for. I didn’t write for you nor for any particular individual, and it’s fair if you are not interested in it. I also added a table of content at the beginning, so you can jump directly to the relevant section (Technical Side) skipping the (in my opinion needed) introduction completely, if you wish. Cheers

    • Brickardo@feddit.nl
      link
      fedilink
      English
      arrow-up
      12
      arrow-down
      7
      ·
      5 months ago

      Tell me you don’t read often without telling me you don’t read often:

    • loudwhisper@infosec.pubOP
      link
      fedilink
      English
      arrow-up
      13
      arrow-down
      2
      ·
      5 months ago

      How do you get this? Anything that tries to force a light mode?

      This is how the site is supposed to look like (there is no light/dark theme selection):

      • ilmagico@lemmy.world
        link
        fedilink
        English
        arrow-up
        12
        ·
        edit-2
        5 months ago

        I was reading the site on Android, and it looked dark, but after seeing this comment, I tried disabling Android system wide dark mode, and sure enough it became white like in the screenshot! For the record, I tried with both Firefox and a Chromium-based browser.

        • loudwhisper@infosec.pubOP
          link
          fedilink
          English
          arrow-up
          15
          ·
          5 months ago

          Thanks! I went and tried on my phone and indeed setting Firefox to light mode indeed causes that horrendous and unreadable result. I will need to figure out way, eventually, and provide an alternative light scheme.

  • ctkatz@lemmy.ml
    link
    fedilink
    English
    arrow-up
    15
    ·
    5 months ago

    there are too many points of failure for me to ever be comfortable using the cloud as a primary storage option.

    i’ve always maintained this opinion when “the cloud” started being touted as being the future. and yet more corporations (including mine) are reliant on it. i mean sure, i can log in on my home computer and have some access to stuff as though i were physically at the office but that convenience ain’t worth the headache if the main storage site crashes.

    • figjam@midwest.social
      link
      fedilink
      English
      arrow-up
      5
      ·
      5 months ago

      there are too many points of failure for me to ever be comfortable using the cloud as a primary storage option.

      If everything that you run is local as in the same physical location and there is no requirement for external or internet access then sure. Not everyone has that luxury. Otherwise, There are the same number of points of failure in a non-cloud configuration. You just feel more comfortable with those because you have direct hands on control.

      • corsicanguppy
        link
        fedilink
        English
        arrow-up
        1
        ·
        5 months ago

        . You just feel more comfortable with those because you have direct hands on control.

        You write “actually following best practice instead of faking it and lying” funny.

        • schizo@forum.uncomfortable.business
          link
          fedilink
          English
          arrow-up
          2
          ·
          5 months ago

          There are places that actually do that?

          Can you provide a list, because I’d like to work there.

          (I do not have 25 years of sysadmin angst over nobody ever doing shit right until after it’s on fire.)

        • figjam@midwest.social
          link
          fedilink
          English
          arrow-up
          2
          ·
          5 months ago

          You write “actually following best practice instead of faking it and lying” funny.

          Are you implying that the various cloud vendors lie about the way they configure their environments or admins don’t have emotional biases or something else entirely?

    • Tja@programming.dev
      link
      fedilink
      English
      arrow-up
      6
      arrow-down
      2
      ·
      5 months ago

      If the storage “crashes” it doesn’t matter if it’s in the cloud or on-prem.

      With the cloud you get two substantial advantages:

      • the storage is built so it doesn’t break so easily. I trust AWS engineers more than Mike, no matter how cool Mike is to hang out with. Additionally, if the storage breaks while Mike is on vacation we’re screwed, with the cloud you get a whole team 24/7 on it.
      • you can prevent data loss with backups or multi-region setups with a few clicks/terraform lines. Try telling the PO that you need to rent datacenter space in Helsinki and Singapore for redundancy…

      Of course all this costs big bucks, but technically it’s superior, easier and less risky.

      • corsicanguppy
        link
        fedilink
        English
        arrow-up
        7
        arrow-down
        1
        ·
        5 months ago

        trust AWS engineers more than Mike, no matter how cool Mike is

        1. AWS engineers’ first responsibility is to shareholders

        2. Mike’s responsibility is to your same boss.

        They are not the same.

        Bonus: you can see Mike’s certs are real.

        • Tja@programming.dev
          link
          fedilink
          English
          arrow-up
          1
          arrow-down
          1
          ·
          5 months ago

          It’s not about responsibility (and only the c suite reports to the shareholders, not Mike), it’s about capability, visibility, tooling and availability.

      • nexusband@lemmy.world
        link
        fedilink
        English
        arrow-up
        5
        arrow-down
        1
        ·
        5 months ago

        the storage is built so it doesn’t break so easily. I trust AWS engineers more than Mike, no matter how cool Mike is to hang out with. Additionally, if the storage breaks while Mike is on vacation we’re screwed, with the cloud you get a whole team 24/7 on it.

        That’s easily mitigated just following established standards. Redundancy is cheaper than anything else in the aftermath and documentation can be done easy with automation.

        you can prevent data loss with backups or multi-region setups with a few clicks/terraform lines. Try telling the PO that you need to rent datacenter space in Helsinki and Singapore for redundancy…

        You don’t, you rent rack space in a location far enough away but close enough to get the data in a few hours.

        It’s neither superior, easier or less risky, it’s just a shift in responsibility. And in most cases, it’s so expensive that a second or third on site engineer is payed for.

        • Tja@programming.dev
          link
          fedilink
          English
          arrow-up
          1
          arrow-down
          1
          ·
          5 months ago

          And what is simpler and faster, renting rack space in another continent (and buying, shipping, racking and initializing) or editing your terraform file?

          • nexusband@lemmy.world
            link
            fedilink
            English
            arrow-up
            3
            ·
            5 months ago

            Why on another continent? Except maybe VDI, some direct calls to some LLM or some insane scales, there’s nothing really that needs those round trip times.

            • ErrorCode@lemmy.world
              link
              fedilink
              English
              arrow-up
              1
              ·
              5 months ago

              Also data rules / data privacy. Some things need to have the original in Europe; China & Russia also need their data separated from others.

          • loudwhisper@infosec.pubOP
            link
            fedilink
            English
            arrow-up
            2
            ·
            5 months ago

            Not OP, but they are comparable efforts, especially since it’s a relatively infrequent activity. You can rent dedicated boxes with off-the-sheld hardware almost instantly, if you don’t want to deal with the hardware procurement, and often you can do that via APIs as well. And of course both options are much, much, much cheaper than the Cloud solution.

            For sure speed in general is something Cloud provide. I would say it’s a very bad metric though in this context.

            • Tja@programming.dev
              link
              fedilink
              English
              arrow-up
              1
              arrow-down
              1
              ·
              5 months ago

              My last customer (global insurance company) provisions several systems a day. Now moving to hundreds via Jenkins. Frequency is environment dependent.

              • loudwhisper@infosec.pubOP
                link
                fedilink
                English
                arrow-up
                1
                ·
                5 months ago

                If your compute needs expand that much everyday, and possibly shrink in others, than your use-case is one that can benefit from Cloud (I covered this in the post).

                That said, if provisioning means recycle, then it’s obviously not a problem.

                This is a very rare requirement. Most companies’ load is fairly stable and relatively predictable, which means that with a proper capacity planning, increasing compute resources is something that happens rarely too. So rarely that even a lead time for hardware is acceptable.

                So if I may ask (and you can tell), what is the purpose of provisioning that many systems each day? Are they continuously expanding?

                • Tja@programming.dev
                  link
                  fedilink
                  English
                  arrow-up
                  1
                  ·
                  edit-2
                  5 months ago

                  Agree to disagree. Banking, telecommunications, insurance, automotive, retail are all industries where I have seen wild load fluctuations. The only applications where I have seen constant load are simulations: weather, oil&gas, scientific. That’s where it makes sense to deploy your own hardware. For all else, server less or elastic provisioning makes economic sense.

                  Edit to answer the last question: to test variable loads, in the last one. Imagine a hurricane comes around and they have to recalculate a bunch of risk components. But can be as simple as running CI/CD tests.

    • IphtashuFitz@lemmy.world
      link
      fedilink
      English
      arrow-up
      4
      ·
      5 months ago

      Having done everything from building my own servers 30 years ago to managing hundreds of servers in data centers to now managing hundreds of instances and other services in AWS, I’ll gladly stick with AWS. The hardware management alone makes it well worth the overhead.

      25 or so years ago I had to troubleshoot a hardware issue in a SCSI-based server with 6 hard drives in it. A drive appeared to be failing so I replaced it and immediately another drive failed, then another, and so on. After almost a full day of troubleshooting later and we realized the power supply was actually the culprit and could no longer provide sufficient power to the full set of hard drives.

      20 years ago while managing 700+ servers in a datacenter we had to manage a recall of about 400 of them thanks to the Capacitor plague that caused a handful of our servers to literally burst into flames.

      Hardware failures like the above and dozens of others were mitigated in most cases thanks to redundancies in the software we wrote. But dealing with hardware failures and the resulting software recovery was a real PITA.

      With AWS I may occasionally have a Linux instance lock up due to a hardware failure but it’s usually fairly easy to reboot the instance and have it migrate to new hardware. It’s also trivial to migrate a server to run on more (or less) number of CPU’s, RAM, etc. with only a couple of minutes of downtime.

      The more advanced services AWS offers like object storage, queues, databases, etc. are even more resilient. We occasionally get notified that a replica for one of these services had failed or was determined to be on hardware that was failing, and it was automatically replaced with a new replica.

      I’d much rather work this way than the way I did 20+ years ago.

      • loudwhisper@infosec.pubOP
        link
        fedilink
        English
        arrow-up
        2
        ·
        5 months ago

        Why not outsourcing just the hardware then? Dedicated servers and Kubernetes slapped on them. Hardware failure mitigated for the most part, and the full effort goes into making the cluster as resilient as possible, for 1/5 of the cost of AWS. If machines burn, it’s not your problem (you can have them spread over multiple sites, DCs, rooms, racks) anymore.

        • IphtashuFitz@lemmy.world
          link
          fedilink
          English
          arrow-up
          3
          ·
          5 months ago

          We did that (with Rackspace) for years before migrating to AWS. AWS is still far better from a service & flexibility perspective.

          My employers website has certain times of the year where we see a huge increase in web traffic. When we had a hosted solution it took weeks of preparation to provision additional web servers to handle that load. We had to submit formal requests for additional servers, document how to wire them into our network & required firewall rules, etc. Then we had to wait an arbitrary number of days for them to do the work. And then we had to repeat that whole process when we no longer needed the additional capacity.

          With AWS we just define an auto scaling group and additional web servers are spun up automatically when demand is high, and frees them up again when no longer needed. Even if we didn’t use auto scaling we could easily automate this sort of thing via terraform or other tools and spin up additional instances in minutes instead of days.

  • iso@lemy.lol
    link
    fedilink
    English
    arrow-up
    9
    ·
    edit-2
    5 months ago

    Yes, I hate cloud too. Now tell this to my company, which received about 100k dollar credits from Azure and Google Cloud :)

        • loudwhisper@infosec.pubOP
          link
          fedilink
          English
          arrow-up
          13
          ·
          5 months ago

          Oh yeah, I know that that’s a thing. It’s a practice not too different from the stereotypical drug dealer who gets you hooked on free drugs. In this case the idea is that if you start there, you get vendor-locked and you will have to pay that amount many times over. I understand the appeal from the company perspective, though.

          • iso@lemy.lol
            link
            fedilink
            English
            arrow-up
            4
            ·
            5 months ago

            Yes absolutely true. For example, GKE looks very nice, but when we use one of their features, it creates the need to use other features too. That’s why I warned the boss a lot. Even though they have great features, we try to use generic applications to avoid hooks.

            I hope they don’t take the credit back :)

  • Lettuce eat lettuce@lemmy.ml
    link
    fedilink
    English
    arrow-up
    8
    ·
    5 months ago

    Very good read. I totally agree with your sentiment that more and more, “engineering” is becoming just gluing together and managing cloud services and features.

    My job as a sys admin has become the same. It’s not about actually understanding the technology at a deep level and troubleshooting problems, it’s about learning specific applets and features to click on and running down daily and weekly checklists.

    • loudwhisper@infosec.pubOP
      link
      fedilink
      English
      arrow-up
      5
      ·
      5 months ago

      I feel you very much. Security work is also somewhat similar.

      I think this takes a way basically the component that made it interesting, understanding what you are doing to the point that you can build stuff.

      it’s about learning specific applets and features to click on and running down daily and weekly checklists.

      Well said.

    • rottingleaf@lemmy.zip
      link
      fedilink
      English
      arrow-up
      1
      ·
      edit-2
      5 months ago

      “engineering” is becoming just gluing together and managing cloud services and features.

      Temporarily becoming.

      Just like China had some social and cultural changes since being closed and till the Opium wars.

      Systems are built around people and limited by what a human can conceive and make work. We don’t evolve that fast.

      Also dependency on big centers has led to catastrophes in the past and will lead to those again.

      It will all crash with a huge bang.

      I’m confident of this, anyone who wants may call me a luddite.

      • Lettuce eat lettuce@lemmy.ml
        link
        fedilink
        English
        arrow-up
        2
        ·
        5 months ago

        Let’s hope that people will start to favor on-prem solutions and smaller independent cloud providers vs the massive trillion dollar corpo clouds that control so much now.

    • Tja@programming.dev
      link
      fedilink
      English
      arrow-up
      3
      arrow-down
      5
      ·
      5 months ago

      And that’s a good thing, IMHO. As an architect I don’t want to rely on some single genius knowing secret incantations or anything like that.

      Boring, tried and true services, repeatedly put together and if the organization allows the time for it, with excessive documentation.

      • conciselyverbose@sh.itjust.works
        link
        fedilink
        English
        arrow-up
        5
        ·
        5 months ago

        No one’s talking about secret incantations.

        They’re talking about knowing how your applications actually work, so you’re not tied to the whims of a third party.

              • conciselyverbose@sh.itjust.works
                link
                fedilink
                English
                arrow-up
                1
                ·
                edit-2
                5 months ago

                That’s literally the entire chain you clicked down.

                The fact that cloud provider calls aren’t based in any kind of core principles and force you to spend all your resources understanding their nonsensical structure instead of what your code actually does.

                • Tja@programming.dev
                  link
                  fedilink
                  English
                  arrow-up
                  1
                  arrow-down
                  2
                  ·
                  5 months ago

                  Wrong. You don’t know how it’s implemented, but you very much know what they do. Even heard about abstraction?

      • Lettuce eat lettuce@lemmy.ml
        link
        fedilink
        English
        arrow-up
        5
        arrow-down
        1
        ·
        5 months ago

        Straw man. I’m encountering sys admins and systems “engineers” who don’t know how to spec out a server, don’t understand how certificates work, don’t understand basic IP addressing principles, don’t understand basic networking topology.

        They just know how to click a list of specific buttons in a GUI for one specific Corpo vendor.

        Maybe that is fine for a Jr. Admin just starting out, but it isn’t what you want for the folks in charge of building, upgrading, and maintaining your company’s infrastructure.

        There’s nothing wrong with making interfaces simpler and easier to understand. And there’s nothing wrong with building simplified abstractions on top of your systems to gain efficiency. But this should not be done at the cost of actual deep understanding and functionality.

        The people you call when things go badly wrong will always be the folks that have that deep understanding and competency. It already has started hitting the developer community in the last few years. The Jr. Devs that did a 3 month boot camp where they learned nothing but how to parrot code and slap APIs together, are getting laid off and cannot find work.

        The devs that went to school for Comp Sci, that have years of real world experience, and actually understand the theory and the nuts and bolts of the underlying tech, they are still largely employed and have little trouble finding work.

        I think the same will happen soon in the IT world. Deep knowledge and years of dirty, greasy hands will always be desirable over a parrot that only knows how to click GUI buttons in a specific order.

      • loudwhisper@infosec.pubOP
        link
        fedilink
        English
        arrow-up
        3
        ·
        5 months ago

        Is that what you get with Cloud? Because there are still a million ways to shoot yourself in the foot. The main difference is that the single genius doesn’t need to implement things him/herself, but decisions still need to be taken and fragile setups can still be built.

        Imagine an ec2 instance in a satellite account performing some business critical function with an instance role, whose custom IAM policy allows to do it in another account. Clouds are not giving you good engineering, they are giving you premade building blocks, you can absolutely still make a mess with those. Even more, the complexity and the immense portfolio of features can allow very creative ways to build very low-quality systems.

        I think you can have good, boring, simple systems built by engineers. With or without Cloud services.

        • Tja@programming.dev
          link
          fedilink
          English
          arrow-up
          1
          ·
          5 months ago

          You can still make a mess, but you can’t fuck up the building blocks, so it’s a big improvement.

          Using an ec2 instance is already a yellow flag, you have higher level services for most tasks.

          • loudwhisper@infosec.pubOP
            link
            fedilink
            English
            arrow-up
            1
            ·
            5 months ago

            Yeah in general you can’t mess the building blocks from the PoV of availability or internal design. That is true, since you are outsourcing it. You can still mess them up from other points of view (think about how many companies got breached due to misconfigured S3 buckets).

  • Jackhammer_Joe@lemmy.world
    link
    fedilink
    English
    arrow-up
    8
    arrow-down
    1
    ·
    5 months ago

    I’m immensely disappointed!

    Not kidding: when I first saw the post title, I was fully convinced that I’ll read the post of a crazy person, rambling about (rain) clouds.

    • loudwhisper@infosec.pubOP
      link
      fedilink
      English
      arrow-up
      5
      ·
      5 months ago

      I am sorry! As an amateur landscape photographer I actually like very much those clouds. There are a few r-word posts about people hating those clouds though, but I checked and they are nowhere near as long as you would expect a proper rant to be

  • TechNerdWizard42@lemmy.world
    link
    fedilink
    English
    arrow-up
    8
    arrow-down
    2
    ·
    5 months ago

    Anything that requires a fancy buzzword is usually stupid but a good way to make money for someone. The “cloud” has always existed as offsite hosting. Off-site shared servers, VPSs, whatever. It’s no different than running CPanel on an LAMP VPS in 2003.

    But calling it “the cloud” gave all the business majors a hard on and then the accounts department realized they could manipulate share pricing by reducing the amount of assets a company holds. It’s the same stupid reason many companies don’t own their corporate headquarters or remote centers. They lease the, even if from themselves through another holding. It looks better on paper so the share price goes up. It’s all mind boggling stupid.

    • MSids@lemmy.world
      link
      fedilink
      English
      arrow-up
      8
      arrow-down
      1
      ·
      5 months ago

      The cloud today significantly different than the 2003 cpanel LAMP server. It’s a whole new landscape. Complex, highly-available architectures that cannot be replicated in an on-prem environment are easily built from code in minutes on AWS.

      Those capabilities come with a steep learning curve on how to operate them in a secure and effective manor, but that’s always going to be the case in this industry. The people that can grow and learn will.

      • TechNerdWizard42@lemmy.world
        link
        fedilink
        English
        arrow-up
        6
        ·
        5 months ago

        I’m fully aware of the few buzzword and marketing pitches that cloud hosting uses. I’m forced to use both GCP and AWS for different contracts and I’m good at it.

        The real truth is that most websites and internet services do not need scale. They do not need all this crap. A Pentium 3 could host all the data for most of these businesses and services. You don’t need serverless lambda functions to handle an api when an actual endpoint does the same thing to pull some info. The few companies that need such distributed computing and power, will need a big on-site or off-site implementation. It makes sense for that sometimes. But most times, it doesn’t even then. You’re just outsourcing your engineering and paying a premium.

        I have seen so many startups spin up cloud accounts costing thousands of dollars a month when they’re in “private beta stealth”. Literally a $500 laptop could host all of their services just as quickly with no monthly fee. But as long as the VCs are paying, just flush that cash down.

        • MSids@lemmy.world
          link
          fedilink
          English
          arrow-up
          3
          arrow-down
          1
          ·
          edit-2
          5 months ago

          The costs are definitely a huge consideration and need to be optimized. A few years back we ran a POC of Open Shift in AWS that seemed to idle at like $3k/mo with barely anything running at all. That was a bad experiment. I could compare that to our new VMWare bill, which more than doubled this year following the Broadcom acquisition.

          The products in AWS simplify costs into an opex model unlike anything that exists on prem and eliminate costly and time consuming hardware replacements. We just put in new load balancers recently because our previous ones were going EoL. They were a special model that ran us a about a half-mil for a few HA pairs including the pro services for installation assistance. How long will it take us to hit that amount using ALBs in AWS? What is the cost of the months that it took us to select the hardware, order, wait 90 days for delivery, rack-power-connect, configure with pro services, load hundreds of certs, gather testers, and run cutover meetings? What about the time spent patching for vulnerabilities? In 5-7 years it’ll be the same thing all over again.

          Now think about having to do all of the above for routers, switches, firewalls, VM infra, storage, HVAC, carrier circuits, power, fire suppression.