• elucubra@sopuli.xyz
    link
    fedilink
    English
    arrow-up
    22
    arrow-down
    12
    ·
    1 day ago

    I’m trying out Google’s Gemma4 LLM, which is run locally, and is touted as a 100% private model.

    Asking it some questions about itself, at one point it acknowledged that chats were sent to “developers”.

    • nightlily@leminal.space
      link
      fedilink
      English
      arrow-up
      11
      ·
      15 hours ago

      You mean it hallucinated a positive response to your leading question as it is meant to? You are operating on a fundamental misunderstanding of what LLMs do. Even if what you said is true, an LLM would have no knowledge of that unless it was explicitly told as such as an input - and why would they be stupid enough to do that?

      • elucubra@sopuli.xyz
        link
        fedilink
        English
        arrow-up
        1
        arrow-down
        2
        ·
        edit-2
        5 hours ago

        You are welcome to try. I can pastebin the prompt. I asked it about itself, the model. It replied that it didn’t exist. I pointed it the the docs, from the Google page. It acknowledged the page was legit, and told me there was no mention of Gemma 4, although there were like 20 mentions, including download links. It insisted. It took me pointing out the specific paragraphs to have it say "this may indicate there is Gemma 4 model. May be…

        At some point it told me I was hallucinating.

        • nightlily@leminal.space
          link
          fedilink
          English
          arrow-up
          2
          ·
          2 hours ago

          I don’t need to try. You aren’t learning facts from interrogating an LLM. If it doesn’t have information, it will make up a result. If it does have information, it will make up a result. Even that is personifying it too much because really the transformer has no concept of what „making something up“ is. It takes an input and gives an output, no matter what.

    • natebluehooves@pawb.social
      link
      fedilink
      English
      arrow-up
      19
      arrow-down
      1
      ·
      1 day ago

      llama.cpp doesn’t have the ability to send telemetry because the next word predictor says so. you can confirm with wireshark.

      • elucubra@sopuli.xyz
        link
        fedilink
        English
        arrow-up
        1
        ·
        5 hours ago

        Oh, I don’t ask for actual answers, but asking it to provide bibliography often points me to the sources, so that I can draw my own conclusions.

    • hansolo@lemmy.today
      link
      fedilink
      English
      arrow-up
      11
      arrow-down
      1
      ·
      1 day ago

      Did the LLM tell you it’s 100% private?

      What else did the LLM tell you?