• Novis@lemdro.id
    link
    fedilink
    English
    arrow-up
    48
    ·
    2 days ago

    This is problematic because anything on your web pages might now influence unrelated answers. You could have outdated information on some forgotten page, or contradictory details across different sections. Google’s AI might grab any of this and present it as the answer. If you allow user-generated content anywhere on your site (like forum posts or comments), someone could post fake support contact info, and Google might surface that to users searching for how to contact your company. Now scammers have a direct route to your customers.

    OH FUUUUUUN

    • ftmpch@lemmy.world
      link
      fedilink
      English
      arrow-up
      2
      arrow-down
      4
      ·
      1 day ago

      You could have outdated information on some forgotten page, or contradictory details across different sections […] If you allow user-generated content anywhere on your site (like forum posts or comments), someone could post fake support contact info,

      None of those things would be Google’s fault, would they?

      • pinball_wizard@lemmy.zip
        link
        fedilink
        English
        arrow-up
        7
        ·
        1 day ago

        Google had the ability to cleanup the data they present, but I read that they stripped out quality checking, because they realized that people spend more time searching (and looking at their ads) when the results are shittier.

      • Novis@lemdro.id
        link
        fedilink
        English
        arrow-up
        1
        ·
        1 day ago

        The problem here is that no one that makes these LLM/AI/whatever are doing ENOUGH DUE DILIGENCE to make sure the data that they’re scrapping is good and accurate to improve the AI’s output. This has been an issue since the beginning and with how much data they’re taking, there’s no good way to get it to 100% accurate. And there was a study put out last year that said it doesn’t take much bad info to poison the AI output. And this is the stuff that these big tech companies are trying to force us all to use in our day-to-day. ALSO YEAH it is Google’s fault because it’s their dog. Their dog is taking the data without understanding WHAT the data they’re taking is, they trained the dog, they have the responsibility for what the dog does out on the internet. It the dog is leading people off a cliff, that’s on Google.

  • pinball_wizard@lemmy.zip
    link
    fedilink
    English
    arrow-up
    3
    ·
    edit-2
    17 hours ago

    I sure hope so. Google doesn’t benefit me, so why should my website benefit Google.

    I have a few things sitting out there for the AI to train on, which no human will ever encounter.

        • omarabid@lemmy.worldOP
          link
          fedilink
          English
          arrow-up
          2
          ·
          2 days ago

          I disabled all of Cloudflare “features” because it’s wrong most of the time. I sent you a message, would be interested to see why this is happening.

          • Pamasich@kbin.earth
            link
            fedilink
            arrow-up
            2
            ·
            1 day ago

            So I’m currently at work, where we use Fortinet, which is blocking your website for “Phishing”.

            Before that, your website made my browser ring the alarm with the error “net::ERR_CERT_AUTHORITY_INVALID”.

            You might want to look into those two.

            • omarabid@lemmy.worldOP
              link
              fedilink
              English
              arrow-up
              2
              ·
              1 day ago

              The certificate issue is weird since it’s handled by Cloudflare and I can see a valid one.

              Th fortinet one. I had a similar issue with a different ISP in the US. I am not sure what’s triggering their firewalls/systems but one guess is that it’s dumb and interpreting codeinput as something that has to do with hacking.