• xavier666@lemm.ee
    link
    fedilink
    English
    arrow-up
    6
    ·
    15 days ago

    My company is realizing that hosting a model which will be private, cost-effective, and performing better than traditional algorithms is like finding a unicorn. Few months back, the top execs were jumping around GenAI like a bunch of kids. Fortunately, the Sr. research head beat some sense into them.

    • falkerie71@sh.itjust.works
      link
      fedilink
      English
      arrow-up
      2
      ·
      15 days ago

      You’re lucky there’s a higher up that could talk down the even higher ups. Though, sometimes it’s not even about the r&d teams.

      I saw company wide HR educational emails or courses telling you how to improve you work quality/efficiency, and one of them tells us to “research AI” and learn how to utilize it, talking about how great it is and improved the work efficiency by 30%. Sure, it has its uses, but I won’t go touting how great it is. And with how ChatGPT works, you have to be the biggest idiot in the world to upload all your sensitive stuff to ChatGPT just for it to make a spreadsheet faster. But without these disclaimers in the email, I doubt regular clerical staff knows about this, and it’s extremely dangerous.

    • Zos_Kia@lemmynsfw.com
      link
      fedilink
      English
      arrow-up
      1
      ·
      15 days ago

      What kind of use-cases was it, where you didn’t find suitable local models to work with ? I’ve found that general “chatbot” things are hit and miss but more domain-constrained tasks (such as extracting structured entities from unstructured text) are pretty reliable even on smaller models. I’m not counting my chickens yet as my dataset is still somewhat small but preliminary testing has been very promising in that regard.

      • xavier666@lemm.ee
        link
        fedilink
        English
        arrow-up
        2
        ·
        15 days ago

        What kind of use-cases was it, where you didn’t find suitable local models to work with ?

        Any time you ask very domain specific questions; eg “i have collected some soil samples from the mesolithic age near the Amazon basin which have high sulfur and phosphorus content compared to my other samples. What factors could contribute to this distribution?”, both of-the-shelf local models & OpenAI fail.

        The main reason is because these models are not trained on highly-specialized domains of text. Sometimes the models start hallucinating and which reduces our trust upon them.

        • Zos_Kia@lemmynsfw.com
          link
          fedilink
          English
          arrow-up
          2
          ·
          15 days ago

          “i have collected some soil samples from the mesolithic age near the Amazon basin which have high sulfur and phosphorus content compared to my other samples. What factors could contribute to this distribution?”

          Haha yeah the top execs were tripping balls if they thought some off-the-shelf product would be able to answer this kind of expert questions. That’s like trying to replace an expert craftsman with a 3D printer.