• lanolinoil@lemmy.world
    link
    fedilink
    English
    arrow-up
    62
    arrow-down
    27
    ·
    1 month ago

    It will design the machines to build the autonomous robots that mine the cobalt… doing the jobs of several companies at one time and either freeing up several people to pursue leisure or the arts or starve to death from being abandoned by society.

    • riodoro1@lemmy.world
      link
      fedilink
      English
      arrow-up
      73
      arrow-down
      3
      ·
      1 month ago

      Have you seen the real fucking world?

      It’s gonna make the rich richer and the poor poorer. At least until the gilded age passes.

    • funkless_eck@sh.itjust.works
      link
      fedilink
      English
      arrow-up
      19
      arrow-down
      8
      ·
      1 month ago

      AI absolutely will not design machines.

      It may be used within strict parameters to improve the speed of theoretically testing types of bearing or hinge or alloys or something to predict which ones would perform best under stress testing - prior to acutal testing to eliminate low-hanging fruit, but it will absolutely not generate a new idea for a machine because it can’t generate new ideas.

      • lanolinoil@lemmy.world
        link
        fedilink
        English
        arrow-up
        13
        arrow-down
        10
        ·
        1 month ago

        The model T will absolutely not replace horse drawn carts – Maybe some small group of people or a family for a vacation but we’ve been using carts to do war logistics for 1000s of years. You think some shaped metal put together is going to replace 1000s of men and horses? lol yeah right

        • funkless_eck@sh.itjust.works
          link
          fedilink
          English
          arrow-up
          7
          arrow-down
          5
          ·
          1 month ago

          apples and oranges.

          You’re comparing two products with the same value prop: transporting people and goods more effectively than carrying/walking.

          In terms of mining, a drilling machine is more effective than a pickaxe. But we’re comparing current drilling machines to potential drilling machines, so the actual comparison would be:

          • is an AI-designed drilling machine likely to be more productive (for any given definition of productivity) than a human-designed one?

          Well, we know from experience that when (loosely defined) “AI” is used in, for e.g. pharma research, it reaps some benefits - but does not replace wholesale the drug approval process and its still a tool used by - as I originally said - human beings that impose strict parameters on both input and output as part of a larger product and method.

          Back to your example: could a series of algorithmic steps - without any human intervention - provide a better car than any modern car designers? As it stands, no, nor is it on the horizon. Can it be used to spin through 4 million slight variations in hood ornaments and return the top 250 in terms of wind resistance? Maybe, and only if a human operator sets up the experiment correctly.

          • lanolinoil@lemmy.world
            link
            fedilink
            English
            arrow-up
            8
            arrow-down
            1
            ·
            edit-2
            1 month ago

            No, the thing I’m comparing is our inability to discern where a new technology will lead and our history of smirking at things like books, cars, the internet and email, AI, etc.

            The first steam engines pulling coal out of the ground were so inefficient they wouldn’t make sense for any use case than working to get the fuel that powers them. You could definitely smirk and laugh about engines vs 10k men and be totally right in that moment, and people were.

            The more history you learn though, you more you realize this is not only a hubrisy thing, it’s also futile as how we feel about the proliferation of technology has never had an impact on that technology’s proliferation.

            And, to be clear, I’m not saying no humans will work or have anything to do – I’m saying significantly MORE humans will have nothing to do. Sure you still need all kinds of people even if the robots design and build themselves mostly, but it would be an order of magnitude less than the people needed otherwise.

            • Beetlejuice001@lemmy.wtf
              link
              fedilink
              English
              arrow-up
              1
              ·
              1 month ago

              Maybe I’m pessimistic but all I see is every call center representative disappearing and that’ll be it

          • sailingbythelee@lemmy.world
            link
            fedilink
            English
            arrow-up
            5
            ·
            1 month ago

            I agree that AI is just a tool, and it excels in areas where an algorithmic approach can yield good results. A human still has to give it the goal and the parameters.

            What’s fascinating about AI, though, is how far we can push the algorithmic approach in the real world. Fighter pilots will say that a machine can never replace a highly-trained human pilot, and it is true that humans do some things better right now. However, AI opens up new tactics. For example, it is virtually certain that AI-controlled drone swarms will become a favored tactic in many circumstances where we currently use human pilots. We still need a human in the loop to set the goal and the parameters. However, even much of that may become automated and abstracted as humans come to rely on AI for target search and acquisition. The pace of battle will also accelerate and the electronic warfare environment will become more saturated, meaning that we will probably also have to turn over a significant amount of decision-making to semi-autonomous AI that humans do not directly control at all times.

            In other words, I think that the line between dumb tool and autonomous machine is very blurry, but the trend is toward more autonomous AI combined with robotics. In the car design example you give, I think that eventually AI will be able to design a better car on its own using an algorithmic approach. Once it can test 4 million hood ornament variations, it can also model body aerodynamics, fuel efficiency, and any other trait that we tell it is desirable. A sufficiently powerful AI will be able to take those initial parameters and automate the process of optimizing them until it eventually spits out an objectively better design. Yes, a human is in the loop initially to design the experiment and provide parameters, but AI uses the output of each experiment to train itself and automate the design of the next experiment, and the next, ad infinitum. Right now we are in the very early stages of AI, and each AI experiment is discrete. We still have to check its output to make sure it is sensible and combine it with other output or tools to yield useable results. We are the mind guiding our discrete AI tools. But over a few more decades, a slow transition to more autonomy is inevitable.

            A few decades ago, if you had asked which tasks an AI would NOT be able to perform well in the future, the answers almost certainly would have been human creative endeavors like writing, painting, and music. And yet, those are the very areas where AI is making incredible progress. Already, AI can draw better, write better, and compose better music than the vast, vast majority of people, and we are just at the beginning of this revolution.

        • funkless_eck@sh.itjust.works
          link
          fedilink
          English
          arrow-up
          5
          arrow-down
          1
          ·
          1 month ago

          can

          might

          sure. But, like I said, those are subject to a lot of caveats - that humans have to set the experiments up to ask the right questions to get those answers.

          • essteeyou@lemmy.world
            link
            fedilink
            English
            arrow-up
            4
            arrow-down
            3
            ·
            1 month ago

            That’s how it currently is, but I’d be astounded if it didn’t progress quickly from now.

            • FiniteBanjo@lemmy.today
              link
              fedilink
              English
              arrow-up
              3
              arrow-down
              1
              ·
              1 month ago

              OpenAI themselves have made it very clear that scaling up their models have diminishing returns and that they’re incapable of moving forward without entirely new models being invented by humans. A short while ago they proclaimed that they could possibly make an AGI if they got several Trillions of USD in investment.

              • essteeyou@lemmy.world
                link
                fedilink
                English
                arrow-up
                3
                arrow-down
                2
                ·
                1 month ago

                5 years ago I don’t think most people thought ChatGPT was possible, or StableDiffusion/MidJourney/etc.

                We’re in an era of insane technological advancement, and I don’t think it’ll slow down.

                • FiniteBanjo@lemmy.today
                  link
                  fedilink
                  English
                  arrow-up
                  4
                  arrow-down
                  1
                  ·
                  edit-2
                  1 month ago

                  Okay but the people who made the advancements are telling you it has already slowed down. Why don’t you understand that? A flawed Chatbot and some art theft machines who can’t draw hands aren’t exactly worldchanging, either, tbh.

                  • essteeyou@lemmy.world
                    link
                    fedilink
                    English
                    arrow-up
                    1
                    arrow-down
                    1
                    ·
                    1 month ago

                    There are other people in the world. Some of them are inventing completely new ways of doing things, and one of those ways could lead to a major breakthrough. I’m not saying a GPT LLM is going to solve the problem, I’m saying AI will.

                  • lanolinoil@lemmy.world
                    link
                    fedilink
                    English
                    arrow-up
                    1
                    arrow-down
                    2
                    ·
                    1 month ago

                    This is such a rich-country-centric view that I can’t stand. LLMs have already given the world maybe it’s greatest gift ever – access to a teacher.

                    Think of the 800 million poor children in the world and their access to a Kahn academy level teacher on any subject imaginable with a cellphone/computer as all they need. How could that not have value and is pearl clutching drawing skills becoming devalued really all you can think about it?

            • funkless_eck@sh.itjust.works
              link
              fedilink
              English
              arrow-up
              1
              ·
              edit-2
              1 month ago

              i would be extremely surprised if before 2100 we see AI that has no human operator and no data scientist team even at a 3rd party distributor - and those things are neither a lie, nor a weaselly marketing stunt (“technically the operators are contractors and not employed by the company” etc).

              We invented the printing press 584 years ago, it still requires a team of human operators.

              • essteeyou@lemmy.world
                link
                fedilink
                English
                arrow-up
                1
                ·
                1 month ago

                A printing press is not a technology with intelligence. It’s like saying we still have to manually operate knives… of course we do.

                • funkless_eck@sh.itjust.works
                  link
                  fedilink
                  English
                  arrow-up
                  1
                  arrow-down
                  1
                  ·
                  edit-2
                  1 month ago

                  the comment I originally replied to claimed AI will design the autonomous machines.

                  It will not. It will facilitate some of the research done by humans to aid in the designing of willfully human operated machinery.

                  To my knowledge the only autonomous machine that exists is a roomba, which moves blindly around until it physically strikes an object, rotates a random degree and continues in a new direction until it hits something else.

                  Even then, it is controlled with an app and on more expensive models, some boundary setting.

                  It is extremely generous to call that “autonomy.”

                  • essteeyou@lemmy.world
                    link
                    fedilink
                    English
                    arrow-up
                    1
                    ·
                    1 month ago

                    I was in a self-driving taxi yesterday. It didn’t need to bump into things to figure out where it was.

    • ObliviousEnlightenment@lemmy.world
      link
      fedilink
      English
      arrow-up
      5
      ·
      1 month ago

      either freeing up several people to pursue leisure or the arts or starve to death from being abandoned by society.

      You know EXACTLY which one it’s gonna be.

    • afraid_of_zombies@lemmy.world
      link
      fedilink
      English
      arrow-up
      2
      ·
      1 month ago

      It isn’t the intelligence of the machine designer that is the issue, it is the middlemen and the end user.

      Continuously having to downgrade machines. Wouldn’t want some sales rep seeing something new.

    • Grandwolf319@sh.itjust.works
      link
      fedilink
      English
      arrow-up
      4
      arrow-down
      3
      ·
      1 month ago

      Hahaha, current ML is basically good guessing, that doesn’t really transfer to building machines that actually have to obey the laws of physics.

      • lanolinoil@lemmy.world
        link
        fedilink
        English
        arrow-up
        1
        arrow-down
        1
        ·
        1 month ago

        is it good guessing that you know when you step out of your bed without looking you won’t fall to your death?

        • Mazoku@lemmy.ml
          link
          fedilink
          English
          arrow-up
          18
          arrow-down
          4
          ·
          1 month ago

          Work a blue collar job your whole life and tell me it’s possible. Machines suck ass. They either need constant supervision, repairs all the time, or straight up don’t function properly. Tech bros always forget about the people who actually keep the world chugging.

          • afraid_of_zombies@lemmy.world
            link
            fedilink
            English
            arrow-up
            1
            ·
            1 month ago

            They suck because your employer wouldn’t pay me more for a better machine. Chemical is where it is at, outside of powerplants and some of the bigger pharms the chemical operator is a dead profession. Entire plants are automated with the only people doing work are doing repairs or sales.

        • leftzero@lemmynsfw.com
          link
          fedilink
          English
          arrow-up
          18
          arrow-down
          4
          ·
          1 month ago

          LLMs aren’t going to be designing anything; they’re just fancy auto complete engines with a tendency to hallucinate facts they haven’t been trained on.

          LLMs are preventing real advancements in AI by focusing the attention and funding into what’s evidently a dead end.

                • leftzero@lemmynsfw.com
                  link
                  fedilink
                  English
                  arrow-up
                  2
                  arrow-down
                  1
                  ·
                  edit-2
                  1 month ago

                  LLMs are incapable of “recognising” any patterns they haven’t been trained on.

                  And they don’t really even recognise those, they’re just fancy auto complete engines, simply outputting the highest scored token from their training base based on their input.

                  They’re pattern matching machines; there’s no recognition, inner modelling of new knowledge, self referencing, or understanding of any kind, merely blind statistics.

                  They’re just bigger and fancier Eliza’s, and just as distant as Eliza was from any practical form of intelligence, artificial or natural.

                  While I personally do believe that achieving AGI¹, on a Turing machine is possible, LLMs and how they work are an excellent example in support of John Searle’s arguments against it with his Chinese room though experiment.

                  1— Or at least something equivalent to human intelligence, or better, in the measures by which we consider ourselves to be intelligent, though it’s arguable whether we can really be considered intelligent at all, or we’re just better, more complex, Chinese rooms.

                  • lanolinoil@lemmy.world
                    link
                    fedilink
                    English
                    arrow-up
                    3
                    ·
                    edit-2
                    1 month ago

                    But since we don’t understand how cognition works in living beings almost at all – who’s to say that’s not how ‘actual thinking’ works other than 'I know it when I see it!"