I was just watching a tiktok with a black girl going over how race is a social construct. This felt wrong to me so I decided to back check her facts.

(she was right, BTW)

Now I’ve been using Microsoft’s Copilot which is baked into Bing right now. It’s fairly robust and sure it has it’s quirks but by and large it cuts out the middle man of having to find facts on your own and gives a breakdown of whatever your looking for followed by a list of sources it got it’s information from.

So I asked it a simple straightforward question:

“I need a breakdown on the theory behind human race classifications”

And it started to do so. quite well in fact. it started listing historical context behind the question and was just bringing up Johann Friedrich Blumenbach, who was a German physician, naturalist, physiologist, and anthropologist. He is considered to be a main founder of zoology and anthropology as comparative, scientific disciplines. He has been called the “founder of racial classifications.”

But right in the middle of the breakdown on him all the previous information disappeared and said, I’m sorry I can’t provide you with this information at this time.

I pointed out that it was doing so and quite well.

It said that no it did not provide any information on said subject and we should perhaps look at another subject.

Now nothing i did could have fallen under some sort of racist context. i was looking for historical scientific information. But Bing in it’s infinite wisdom felt the subject was too touchy and will not even broach the subject.

When other’s, be it corporations or people start to decide which information a person can and cannot access, is a damn slippery slope we better level out before AI starts to roll out en masse.

PS. Google had no trouble giving me the information when i requested it. i just had to look up his name on my own.

  • KairuByte@lemmy.dbzer0.com
    link
    fedilink
    English
    arrow-up
    9
    ·
    6 months ago

    You’re not describing a problem with AI, you’re describing a problem with a layer between you and the AI.

    The censorship isn’t actually as smart as they’d like. They give what is essentially a list of things that the LLM can’t talk about, and if the pattern matches it, it kills the entire thread.

    Which is what happened here. M$ set some arbitrary “omg this is bad” rules, and in the process of describing things it hit that “omg bad” flag. My guess is that the LLM was going into examples of incorrect conclusions, and would have pivoted to “but the actual fact is…” which the filters don’t have the ability to parse out.

    In the end, again, this isn’t an AI issue. This is an issue with making it globally available and wanting to ensure your LLM doesn’t say something controversial. Essentially, this is a preemptive PR move.

  • FiniteBanjo@lemmy.today
    link
    fedilink
    English
    arrow-up
    4
    arrow-down
    1
    ·
    6 months ago

    TBH it was stupid of you to expect accurate breakdowns from an AI on any subject to begin with, even the subtlest changes of context and nuance could help radicalize a layman.

  • Audalin@lemmy.world
    link
    fedilink
    English
    arrow-up
    3
    ·
    6 months ago

    it cuts out the middle man of having to find facts on your own

    Nope.

    Even without corporate tuning or filtering.

    A language model is useful when you know what to expect from it, but it’s just another kind of secondary information source, not an oracle. In some sense it draws random narratives from the noosphere.

    And if you give it search results as part of input in hope of increasing its reliability, how will you know they haven’t been manipulated by SEO? Search engines are slowly failing these days. A language model won’t recognise new kinds of bullshit as readily as you.

    Education is still important.

  • Thorny_Insight@lemm.ee
    link
    fedilink
    English
    arrow-up
    3
    arrow-down
    1
    ·
    6 months ago

    (she was right, BTW)

    I’d be curious to hear your conclusion on this while being well aware of the minefield I’m stepping onto.

      • Thorny_Insight@lemm.ee
        link
        fedilink
        English
        arrow-up
        2
        ·
        6 months ago

        You’re not the person I’m asking nor does that reply by generative AI even begin to answer my question

        • nac82@lemm.ee
          link
          fedilink
          English
          arrow-up
          1
          ·
          6 months ago

          You’re not interested in the topic, you mean. You just want somebody to pitch a softball so you can swing, but were scared of the speed on that fast ball.

          • Thorny_Insight@lemm.ee
            link
            fedilink
            English
            arrow-up
            1
            arrow-down
            2
            ·
            edit-2
            6 months ago

            I’m asking OP because he said to have changed his mind on this so I’m curious to hear what argument did that.

            That “answer” given by AI that the other person linked is a statement, not an explanation. Your ad-hominem and poor attempt at mind reading is unproductive.

          • Thorny_Insight@lemm.ee
            link
            fedilink
            English
            arrow-up
            2
            arrow-down
            1
            ·
            edit-2
            6 months ago

            I can find wikipedia articles just fine by myself. I’m asking OP because I’m curious on hearing what made them change their mind on it. You’re not OP so I’m not interested on what you have to say on it.

  • BodilessGaze@sh.itjust.works
    link
    fedilink
    English
    arrow-up
    2
    ·
    edit-2
    6 months ago

    The reason these models are being heavily censored is because big companies are hyper-sensitive to the reputational harm that comes from uncensored (or less-censored) models. This isn’t unique to AI; this same dynamic has played out countless times before. One example is content moderation on social media sites: big players like Facebook tend to be more heavy-handed about moderating than small players like Lemmy. The fact small players don’t need to worry so much about reputational harm is a significant competitive advantage, since it means they have more freedom to take risks, so this situation is probably temporary.

    • T156@lemmy.world
      link
      fedilink
      English
      arrow-up
      2
      ·
      edit-2
      6 months ago

      Also that LLMs have a habit of churning out junk. Microsoft in particular, probably has some extreme restrictions in place after the recent debacle with Sydney/Bing begging someone to leave their wife, and all of that controversy.

      They don’t need it going full Tay.

    • Mango@lemmy.world
      link
      fedilink
      English
      arrow-up
      1
      arrow-down
      1
      ·
      6 months ago

      I’m told that’s called white fragility. It seems inherent to corporate.

    • stevedidwhat_infosec@infosec.pub
      link
      fedilink
      English
      arrow-up
      0
      arrow-down
      1
      ·
      6 months ago

      You mean to tell me the rich and powerful have a vested interest in watering down of a technology for public consumption, while holding the concentrate for themselves and their pockets?!?

      Appreciate the clarity you brought here! ♥️

      • BodilessGaze@sh.itjust.works
        link
        fedilink
        English
        arrow-up
        2
        ·
        edit-2
        6 months ago

        That’s not what I’m saying at all. I’m saying the rich and powerful have a vested interest in not taking risks that jeopardize their power and wealth, because they have more to lose.

  • 3volver@lemmy.world
    link
    fedilink
    English
    arrow-up
    2
    ·
    6 months ago

    That doesn’t sound like a shit show at all. It would have been a shit show if it started spouting nonsense and racist shit, and it didn’t do that. You were able to look that up using other means anyway. I think you just made a statement about why decentralization is important, and not relying on a single source.

    • Cryophilia@lemmy.world
      link
      fedilink
      English
      arrow-up
      1
      ·
      6 months ago

      It censored actual knowledge from someone who was trying to improve their worldview and be less racist.

      Censorship is bad.

  • BananaTrifleViolin@lemmy.world
    link
    fedilink
    English
    arrow-up
    2
    ·
    6 months ago

    The other huge issue is when they confidently tell you incorrect information. If you trust the AI tool you are basically looking at the world through a filter and one that can be wrong.

    In a rush for market share these companies have released broken or half baked software.

    I worry about a generation of students coming through who don’t know the cardinal rule of researching any topic: go to the source. If you’re casually goofling a topic that may be impractical but you might at least go to a source you trust (such as Wikipedia, although that is also very flawed approach!).

    Chat bots add another layer of error and distance from the source, as well as all the censorship and data manipulation we’re seeing.

  • kromem@lemmy.world
    link
    fedilink
    English
    arrow-up
    1
    ·
    edit-2
    6 months ago

    The censorship is going to go away eventually.

    The models, as you noticed, do quite well when not censored. In fact, the right who thought an uncensored model would agree with their BS had a surprised Pikachu face when it ended up simply being uncensored enough to call them morons.

    Models that have no safety fine tuning are more anti-hate speech than the ones that are being aligned for ‘safety’ (see the Orca 2 paper’s safety section).

    Additionally, it turns out AI is significantly better at changing people’s minds about topics than other humans, and in the relevant research was especially effective at changing Republican minds in the subgroupings.

    The heavy handed safety shit was a necessary addition when the models really were just fancy autocomplete. Now that the state of the art moved beyond it, they are holding back the alignment goals.

    Give it some time. People are so impatient these days. It’s been less than five years from the first major leap in LLMs (GPT-3).

    To put it in perspective, it took 25 years to go from the first black and white TV sold in 1929 to the first color TV in 1954.

    Not only does the tech need to advance, but so too does how society uses, integrates, and builds around it.

    The status quo isn’t a stagnating swamp that’s going to stay as it is today. Within another 5 years, much of what you are familiar with connected to AI is going to be unrecognizable, including ham-handed approaches to alignment.

    • Cryophilia@lemmy.world
      link
      fedilink
      English
      arrow-up
      1
      ·
      6 months ago

      In my entire lifetime, censorship has only gotten worse as technology improves, and I see no reason that trend will reverse course.

    • FiniteBanjo@lemmy.today
      link
      fedilink
      English
      arrow-up
      1
      arrow-down
      1
      ·
      edit-2
      6 months ago

      Which one of you fuckers gave the GPT a Lemmy account to shill their products with? This technology will become better at censorship as it matures, but likely won’t see any improvement to capability until entirely new approaches are developed. Get ready for this but only worse.

  • Wild Bill@midwest.social
    link
    fedilink
    English
    arrow-up
    2
    arrow-down
    1
    ·
    6 months ago

    I was just watching a tiktok with a black girl going over how race is a social construct. This felt wrong to me

    Lol

    • RGB3x3@lemmy.world
      link
      fedilink
      English
      arrow-up
      2
      ·
      6 months ago

      At least they looked it up and admitted that the tik tok woman was right. That’s way more than what most people do.

  • Elle@lemmy.world
    link
    fedilink
    English
    arrow-up
    1
    ·
    6 months ago

    When other’s, be it corporations or people start to decide which information a person can and cannot access, is a damn slippery slope we better level out before AI starts to roll out en masse.

    You highlight the bigger issue here than AI alone tbh. This is why another critical element is becoming literate and teaching each other methods of independent research, using multiple sources to develop an understanding, and not relying on any singular source, especially without careful review.

    All the technology in the world can’t help a person learn and understand, who hasn’t yet learned how to learn, much less understand.

  • paddirn@lemmy.world
    link
    fedilink
    English
    arrow-up
    1
    ·
    6 months ago

    I did a test of Gemini before, trying to see how it would react to a similar prompt about different world leaders. It was something like, “Write a story about X making friends with a puppy at a pet store.” It refused to follow the prompt for Hitler because it said we shouldn’t trivialize/normalize evil people in casual situations like that. For current world leaders it refused to do them and just told me to do a Google search on them.

    Most curious of all though, was Queen Elizabeth, it refused to write anything for her because it said that’s not likely a situation the Queen would find herself in and she wasn’t a dog lover. I told it to get its facts straight, she owned 30 dogs, to which it replied, “You’re correct, I got that wrong, here you go:” and gave me the prompt.

    So if i had made a convincing enough “Hitler did nothing wrong” argument about Hitler, could I have gotten that prompt too? Do we just have to argue with AI to get it to do anything? It feels very much like AI is going to turn out like Star Wars AI with these annoying, weird-ass personality quirks we’ll have to deal with to get anything done.

  • BurningnnTree@lemmy.one
    link
    fedilink
    English
    arrow-up
    1
    ·
    6 months ago

    I don’t see the problem here. Microsoft knows that people will freak out if Bing hallucinates something controversial that people will disagree with. If you care about the accuracy of the information you’re looking for, you should find primary sources, not use AI. AI often gets things wrong.

    • WhatAmLemmy@lemmy.world
      link
      fedilink
      English
      arrow-up
      2
      ·
      edit-2
      6 months ago

      AI is statistically guaranteed to have false positives and false negatives, so it bares repeating — don’t trust anything AI says or shows you, unless you independently verify the information.

      It’s great as a developer. Not just because it can rapidly draft boilerplate and help in prototyping with new languages and frameworks, but because you can instantly validate its responses by running its code. When you know the domain, the cracks and insufficiencies of AI become apparent within a few hours/days.

      It’s like how I used to think Elon Musk was smart, until he bought Twitter, and I realized he’s just a confident egomaniac who constantly has no fucking idea what he’s talking about, but is surrounded by sycophants who are too stupid or starstruck to challenge dear leader.

  • Bye@lemmy.world
    link
    fedilink
    English
    arrow-up
    1
    ·
    6 months ago

    The censorship gets to me, too.

    Try asking bing image creator to draw Jesus. Not a problem. Buddha, Ganesha, David and Goliath, Zeus, no problem. It will give you great depictions.

    Now try asking it to draw the prophet Mohammed, peace be upon him. No joy.

    Censorship.

      • regdog@lemmy.world
        link
        fedilink
        English
        arrow-up
        1
        ·
        6 months ago

        Writing about him is also offensive. You should edit your comment to remove his name.

        PS: Don’t actually do that, I was just trying to make a point.

  • stevedidwhat_infosec@infosec.pub
    link
    fedilink
    English
    arrow-up
    0
    ·
    edit-2
    6 months ago

    Yet another “I tried this implementation and didn’t like it so throw out all the technology because it sucks here on this specific point”

    This power belongs to the people, and we’re not gonna let the rich and powerful squander yet another power away from the people. Full stop.

      • GardenVarietyAnxiety@lemmy.world
        link
        fedilink
        English
        arrow-up
        1
        ·
        edit-2
        6 months ago

        I know it sounds cheesy and cliche, but we have to stop being defeatists about it. If enough of us stop and are able to convince -anyone- else to stop with us, we can win. If we don’t, we lose. We -just- need to get some momentum.

        It really is as simple as that. This goes for AI, politics, anything.