• gravitas_deficiency@sh.itjust.works
    link
    fedilink
    English
    arrow-up
    6
    ·
    9 hours ago

    I really don’t think you’re looking at this from the right angle. This isn’t about being lazy. This isn’t about not double checking work.

    My point is that statistically speaking, even the double checkers who check the work of the double checkers may, at some point, miss some really subtle, nuanced condition. Colloquially, these often fall under the category of critical zero-day bugs. Having a language that makes it impossible to code that’s vulnerable to whole categories of exploits and bugs is an objective good. I’m a bit mystified why you’re trying to argue that it’s purely a skill/rigor issue.

    Case in point: the LN-100 inertial nav unit used in the F-22 had a bug in it that caused the whole system to unrecoverably crash as the first squadron flew over the International Date Line as it was being deployed to Kaneda air base in Japan. The only reason why they didn’t have to ditch in the pacific was that the tanker was still in radio range; they had to be shepherded back to Honolulu by the tanker, and Northrop Grumman flew an engineering team out to (very literally, heh) hotfix the planes on the tarmac, and then they continued on to Kaneda without issue. TLDR: even with systems that enforce extreme rigor (code was developed and tested under DO-178B), mistakes can and do happen. Having a language that guards against that is just one more level of safety, and that’s a good thing.

    • ulterno@lemmy.kde.social
      link
      fedilink
      English
      arrow-up
      1
      arrow-down
      6
      ·
      edit-2
      8 hours ago

      Having a language that guards against that is just one more level of safety, and that’s a good thing.

      Yes it is.
      But my point simply is, “caring” about stuff needs to be normalised, instead of over-anti-pedantism and answering concerns with stuff like, “chill dude!”.
      We know very well that not all bugs are memory related.