• @Flumpkin
    link
    English
    04 months ago

    That might actually be the kind of thing where open source AI could help. At least I hope. To detect bias, lies or AI powered filtering / sorting of content.

    • @mods_are_assholes@lemmy.world
      link
      fedilink
      English
      54 months ago

      Ok so this is one of the naive thoughts that makes me upset.

      The open source community can’t even make a distro of linux that is out of the box functional for everyday users and you think somehow they are going to be able to outcompete billion dollar companies that can afford the best gear and devs?

      Look, I bought in heavy to open source early on in the 90s, and have done my best to go open source for every tool I can, but the simple fact is that even the ‘best’ open source projects are severely lacking in aspects and YOU CAN’T TRUST DEVELOPMENT OF AI TO THAT.

      Compare The Gimp to Photoshop. It isn’t even close, why? Because Adobe has a fucktonne of cash to throw at their projects and they have clear direction and motivation.

      I don’t like it

      I’d prefer a fully open source world

      But it isn’t going to happen, and open source AI will always lag behind corporate AI, and considering how fast it has been developing, even being 3 months behind renders a tool useless as an AI detector.

      We aren’t prepared for this and 90% of what everyone on the internet says about AI is poorly informed and full of confabulation, and WORST of all, when you try and explain this to them they get antagonistic.

      We have already seen the threat AI can pose in 2016 with Cambridge Analytica helping to hand trumpty dumpty the election by using AI to focus target vulnerable facebook groups.

      AND THAT AI WAS A FUCKING INFANT compared to what we have now.

      It’s going to be so bad and almost none of you have the slightest clue.

      • @OhNoMoreLemmy@lemmy.ml
        link
        fedilink
        English
        14 months ago

        Honestly, most of what Cambridge analytica did was blackmail, illegal spending, and collusion between campaigns that were legally required to be separate.

        Much of the data processing/ml was intended as a smoke screen to distract from the big stuff that was known to work and consequently legislated against. The problem is that they were so incompetent that the distraction technique was also illegal.

        Maybe the machine learning also worked, but it’s really not clear.

      • @Kedly@lemm.ee
        link
        fedilink
        English
        14 months ago

        See, THIS is the criticism of AI I can actually empathize with, I might even agree with it somewhat