• That’s unprovable without some very strict definitions, but if we take it as a given (and for the record I don’t disagree, so we should) then that’s why the ai isn’t the artist. It’s just a tool an artist could use. MS Paint isn’t an artist either, and like ai neither are many of the people using it, but it still can be used to create art.

        • udon@lemmy.world
          link
          fedilink
          arrow-up
          5
          ·
          1 month ago

          Meh, better approach it to assume it doesn’t understand emotion unless proven otherwise. Does a fork understand what human emotion is? A pillow? You wouldn’t assume that either I guess.

          • Uriel238 [all pronouns]@lemmy.blahaj.zone
            link
            fedilink
            English
            arrow-up
            3
            ·
            1 month ago

            So which of us are p-zombies? We’ve encountered the same problem by suggesting that human beings have consciousness or self awareness, or get what qualia are, except we can’t prove that anyone has any of these things. The difference of AI consciousness within its development community is a sorites paradox. Large AI packages like GPT-4 have more awareness than previous versions, but not as much awareness as humans. But it Chat GPT4 does exceed human control subjects in the Turing test.

            Mind you the Turing is only one of several tests we use to rate how advanced AI is, but we can’t be sure even when an AI can make coffee given a machine and supplies, and construct flat-packed furniture given the IKEA visual instructions, that this counts as AGI, or is sentient.

            Right now, there are artists who use generative AI to create art, and it is as much really art as photography was really art when illustrators were complaining they are just using a machine to replicate a real scene. As much as music production and music synthesis are art.

            Now yes, I get that AI presents risks of workers losing income and their capacity to survive, but every time we toss our sabots into the gearworks to break the machines, we’re kicking overthrow of the system down the line, until we’re where we are today, not only looking at the dissolution of our democracy so that industrialists may continue to exist, but also the destruction of our habitat, because we can’t address what makes them money.

            So capitalism is going to end you either way, unless you end it first. And I expect if you actually tried to make a fortune on your art, you would eventually find yourself selling out all your rights to one of the big corporate controllers, and they would own everything you did, and pay you a pittance for it… Unless you are James Hetfield kind of skilled and lucky. Somehow I doubt you are.

            • udon@lemmy.world
              link
              fedilink
              arrow-up
              3
              ·
              1 month ago

              Welcome to radical constructivism :) The question whether other people or cats can experience emotions is in fact a problem people have been thinking about quite a lot. Answers are not very satisfactory, but one way to think about it (e.g., some constructivists would do that) is that assuming they do have a conscience simplifies your world model. In the case of “AI” though, we have good alternative explanations for their behavior and don’t need to assume they can experience anything.

              The other important bit is that not assuming some phenomenon exists (e.g., “AI” can experience emotions) unless proven otherwise is the basis of modern (positivistic) science.

              • assuming they do have a conscience simplifies your world model.

                Does it? Feels more like it merely excludes them from your model, since your model cannot explain their conscience. If that simplifies your model, then you can apply the same thinking to anything you don’t understand by simply saying it is similar to something else you also can’t explain.

                The other important bit is that not assuming some phenomenon exists (e.g., “AI” can experience emotions) unless proven otherwise

                The problem with this isn’t that it’s literally unprovable, it’s that proving it requires defining “can experience emotions” in a way everyone can agree on. Most trivial definitions that include everything we think ought obviously be included often bring in many things we often think ought be excluded, and many complicated definitions that prune out the things we think ought be excluded, often also cut out things we think should be included

                • udon@lemmy.world
                  link
                  fedilink
                  arrow-up
                  3
                  ·
                  1 month ago

                  Does it?

                  Yes, in the sense that “thing moves around and does stuff” becomes more predictable if you assume a certain degree of consciousness. This is easier than “thing is at this position now, was at a different position before, was at yet another position before that”. You reduce some of the complexity and unpredictability by introducing an explanation for these changes of world state. In my world, so far I worked well with the assumption that other humans and animals have some consciousness and at least I’m not aware of any striking evidence that would raise doubt on that.

                  The problem with this isn’t that it’s literally unprovable

                  Yes, that’s a problem, but it’s relatively similar to the other one. It’s actually quite hard to “prove” anything with real world connection. However, in the case of other humans/animal consciousness, evidence suggests so (at least for me). The evidence in the case of “AI” is different, though. For example, they seem to have no awareness of time and no awareness of the world beyond the limited context of a conversation. Besides a fancy marketing term that suggests there is something similar to living beings involved, what we currently see are admittedly impressive programs that run on statistics, but I don’t need to assume any “consciousness” to explain what they do.

                  • You reduce some of the complexity and unpredictability by introducing an explanation for these changes of world state

                    My concern is that “consciousness” isn’t so much an explanation as it is a sort of heuristic. We feel conscious and have an internal experience, so it seems pretty reasonable to say that such a thing exists, but other than one’s own self, there’s no point where it is certain to exist, and no clear criteria or mechanism that we can point to.

                    What about the p-zombie, the human person who just doesn’t have an internal experience and just had a set of rules, but acts like every other human? What about a cat, who apparently has a less complex internal experience, but seems to act like we’d expect if it has something like that? What about a tick, or a louse? What about a water bear? A tree? A paramecium? A bacteria? A computer program?

                    There’s a continuum one could construct that includes all those things and ranks them by how similar their behaviors are to ours, and calls the things close to us conscious and the things farther away not, but the line is ever going to be fuzzy. There’s no categorical difference that separates one end of the spectrum from the other, it’s just about picking where to put the line.

                    And yes, we have perhaps a better understanding of the mechanism behind how an ai gets from input to output than we do for a human, but it’s not quite a complete one. And the mechanism for how humans get from an input to an output is similarly partially understood. We can see how the arrangement and function of nerve cells in a “brain” lead to the behaviors we see, and have even fully simulated the brains of some organisms with machine code. This is not so dissimilar from how a computational neural network is operated. The categorical difference of “well one is a computer” doesn’t work when we have literally simulated an organic brain also on a computer.

    • Prandom_returns@lemm.ee
      link
      fedilink
      arrow-up
      3
      ·
      1 month ago

      If I ask a painter to paint a landscape, who’s making art, me or the painter?

      Is the painter just a tool?

      • You can’t really have it both ways.

        Is the things just a machine that’s following instructions and synthesizing its training data into different things? Then it’s a tool.

        Is the things making choices and interpreting your inputs to produce a result? Then it’s an artist.

        The painter I buy a commission from is an artist. The ai I use to generate a scene is a tool.

          • I mean, yes?

            That’s very pithy, but the material used as training data was probably produced by artists attempting to create art using tools (ai and otherwise), as well as more mundane data designed and produced by humans with no ai tools and some produced by humans with almost exclusively ai tools.

            • Prandom_returns@lemm.ee
              link
              fedilink
              arrow-up
              2
              ·
              1 month ago

              You probably live in a different world than I do.

              Don’t chicken/egg this. All of the training data was man-made at some point. Until the first LLMs started outputting based on it.

              Secondly, the amount of human-produced content and LLM-produced content that’s in the training data is incomparable. And will continue to be so. Otherwise the models break.