• monotremata@lemmy.ca
    link
    fedilink
    English
    arrow-up
    28
    arrow-down
    6
    ·
    2 days ago

    I mean, not really? Maybe they’re both deep learning neural architectures, but one has been trained on an entire internetful of stolen creative content and the other has been trained on ethically sourced medical data. That’s a pretty significant difference.

    • KeenFlame@feddit.nu
      link
      fedilink
      arrow-up
      13
      ·
      2 days ago

      No, really. Deep learning and transformers etc. was discoveries that allowed for all of the above, just because corporate vc shitheads drag their musty balls in the latest boom abusing the piss out of it and making it uncool, does not mean the technology is a useless scam

      • ILikeTraaaains@lemmy.world
        link
        fedilink
        arrow-up
        7
        ·
        2 days ago

        This.

        I recently attended a congress about technology applied on healthcare.

        There were works that improved diagnosis and interventions with AI, generative mainly used for synthetic data for training.

        However there were also other works that left a bad aftertaste in my mouth, like replacing human interaction between the patient and a specialist with a chatbot in charge of explaining the procedure and answering questions to the patient. Some saw privacy laws as a hindrance and wanted to use any kind of private data.

        Both GenAI, one that improves lives and other that improves profits.

      • monotremata@lemmy.ca
        link
        fedilink
        English
        arrow-up
        3
        arrow-down
        1
        ·
        2 days ago

        Yeah, that’s not what I was disagreeing with. You’re right about that; I’m on record saying that capitalism is our first superintelligence and it’s already misaligned. I’m just saying that it isn’t really meaningless to object to generative AI. Sure the edges of the category are blurry, but all the LLMs and diffusion-based image generators and video generators were unethically trained on massive bodies of stolen data. Seriously, talking about AI as though the architecture is the only significant element when getting good training data is like 90% of the challenge is kind of a pet peeve of mine. And seen in that light there’s a pretty significant distinction between the AI people are objecting to and the AI people aren’t objecting to, and I don’t think it’s a matter of “a meaningless buzzword.”

        • KeenFlame@feddit.nu
          link
          fedilink
          arrow-up
          1
          ·
          14 hours ago

          I totally understand that. The pet peeve of yours, i just disagree with on a fundamental level. The data is the content, and speaking about it as if the data is the technology itself is like talking about clothes in general as being useful or not. It’s meaningless especially if you don’t know about or acknowledge the different types of apparel and their uses. It’s obviously not general knowledge but it would be like bickering about if underwear is a great idea or not, it’s totally up to the individual if they want to wear them, even if being butt naked in public is illegal. If the framework is irrelevant, then the immediate problem isn’t generative AI, especially the perfectly ethical open source models

    • AdrianTheFrog@lemmy.world
      link
      fedilink
      English
      arrow-up
      7
      arrow-down
      1
      ·
      2 days ago

      I think DLSS/FSR/XeSS is a good example of something that is clearly ethical and also clearly generative AI. Can’t really think of many others lol