I thought of this recently (anti llm content within)

The reason a lot of companies/people are obsessed with llms and the like, is that it can solve some of their problems (so they think). The thing I noticed, is a LOT of the things they try to force the LLM to fix, could be solved with relatively simple programming.

Things like better searches (seo destroyed this by design, and kagi is about the only usable search engine with easy access), organization (use a database), document management, etc.

People dont fully understand how it all works, so they try to shoehorn the llm to do the work for them (poorly), while learning nothing of value.

  • 🇰 🌀 🇱 🇦 🇳 🇦 🇰 🇮 @pawb.social
    link
    fedilink
    English
    arrow-up
    6
    arrow-down
    2
    ·
    3 days ago

    And how do you know that the LLM was accurate and gave you the correct information, instead of just making up something entirely novel and telling you what you wanted to hear? Maybe the detail you were searching for could not be found, because it did not actually exist.

    • Blue_Morpho@lemmy.world
      link
      fedilink
      arrow-up
      5
      arrow-down
      2
      ·
      3 days ago

      Maybe the detail you were searching for could not be found, because it did not actually exist.

      He said he clicked the source it quoted.

      Maybe if Google hasn’t been enshittifying search for 10 years, AI search wouldn’t be useful. But I’ve seen the same thing. The forced Gemini summary at the top of Google often has source links that aren’t anywhere on the first page of Google itself.

    • HelloRoot@lemy.lol
      link
      fedilink
      English
      arrow-up
      4
      arrow-down
      5
      ·
      edit-2
      3 days ago

      First, read my text fully before replying.

      But additionally I have a brain and can use it to double check:

      In example 1. I just build it blindly because it’s a game and it doesn’t matter if it’s wrong. But it ended up being correct and I ended up having more fun instead of doing excel for an hour.

      In 2. the math result was not far off from my guesstimate and I confirmed later, it was correct.

      In 3. it gave me a source and I read the source. Google did not lead me to that source.

      When I let LLM write code, I read the code, then I test the code. Here is where I get the most faults. Not in spreadsheets or math or research.

      • Blue_Morpho@lemmy.world
        link
        fedilink
        arrow-up
        4
        arrow-down
        3
        ·
        edit-2
        3 days ago

        It’s weird how there is such a knee jerk hate for a turbo charged word predictor. You’d think there would have been similar mouth frothing at on screen keyboards predicting words.

        I see it as a tool that helps sometimes. It’s like an electric drill and craftsmen are screaming, “BUT YOU COULD DRILL OFF CENTER!!!”

        • jasory@programming.dev
          link
          fedilink
          arrow-up
          3
          arrow-down
          1
          ·
          3 days ago

          The commenter more or less admitted that they have no way of knowing that the algorithm is actually correct.

          In your first analogy it would be like if text predictors pulled words from a thesaurus instead of a list of common words.

          • Blue_Morpho@lemmy.world
            link
            fedilink
            arrow-up
            1
            ·
            3 days ago

            that they have no way of knowing that the algorithm is actually correct.

            He tested it and it was good enough for him. If he wrote the code he’d still not know if it was correct and need to test it. If knowing an algorithm was all that was needed for writing working code, there wouldn’t have been any software bugs in all of computer history until AI.

            text predictors pulled words

            My phone keyboard text predictor lists 3 words and they’re frequently wrong. At best it lists 3 and you have to choose the 1 right word.