• corbin@awful.systems
    link
    fedilink
    English
    arrow-up
    5
    ·
    4 天前

    In practice, the behaviors that the chatbots learn in post-training are FUD and weasel-wording; they appear to not unlearn facts, but to learn so much additional nuance as to bury the facts. The bots perform worse on various standardized tests about the natural world after post-training; there are quantitative downsides to forcing them to adopt any particular etiquette, including speaking like a chud.

    The problem is mostly that the uninformed public will think that the chatbot is knowledgeable and well-spoken because it rattles off the same weak-worded hedges as right-wing pundits, and it’s addressed by the same improvements in education required to counter those pundits.

    Answering your question directly: no, slop machines can’t be countered with more slop machines without drowning us all in slop. A more direct approach will be required.

    • will_a113@lemmy.ml
      link
      fedilink
      English
      arrow-up
      1
      arrow-down
      3
      ·
      4 天前

      Do you have any sources on this? I started looking around for pre-training, training and post-training impact of new input but didn’t find what I was looking for. In just my own experience with retraining (e.g. fine-tuning) pre-trained models, it seems to be pretty easy to add or remove data to get significantly different results than the original model.