Would you participate?

  • Admiral Patrick@dubvee.org
    link
    fedilink
    English
    arrow-up
    23
    ·
    1 day ago

    No. A thousand times “no”. Just seems like a complete waste of resources federating intentional gibberish and more shit for users to block by default.

    Instead, I recommend pushing your instance admins to run something like Nepenthes so that bot traffic is automatically served gibberish in the background instead of actual content. I’ve been doing this for a couple weeks now, and multiple bots are constantly thrashing around in the tarpit.

  • cronenthal@discuss.tchncs.de
    link
    fedilink
    arrow-up
    4
    ·
    22 hours ago

    Hilariously, the industry is doing this job itself. Endless ai generated LinkedIn posts, tweets, reddit comments, news articles etc. will ensure there is increasingly useless data flooding the internet. There is no real way to filter it and it will poison all future models.

  • trollercoaster@sh.itjust.works
    link
    fedilink
    English
    arrow-up
    11
    ·
    1 day ago

    The problem ist that this is all too obvious and can simply be filtered out based on the location. Better would be to deliberately add all sorts of gibberish to regular posts in order to poison machine learning models.

  • Grimy@lemmy.world
    link
    fedilink
    arrow-up
    5
    ·
    edit-2
    1 day ago

    It’s too easy to actually poison an LLM. They aren’t scrapping the web like they used to anymore. Even if they did, they would have filters to pick up on gibberish.